#!/bin/bash
-REPO_URL_BLOB=$(cat /opt/config/repo_url_blob.txt)
-REPO_URL_ARTIFACTS=$(cat /opt/config/repo_url_artifacts.txt)
-DEMO_ARTIFACTS_VERSION=$(cat /opt/config/demo_artifacts_version.txt)
-INSTALL_SCRIPT_VERSION=$(cat /opt/config/install_script_version.txt)
+SCRIPT_VERSION=$(cat /opt/config/script_version.txt)
CLOUD_ENV=$(cat /opt/config/cloud_env.txt)
+MUX_GW_IP=$(cat /opt/config/mux_gw_net_ipaddr.txt)
+MUX_GW_CIDR=$(cat /opt/config/mux_gw_net_cidr.txt)
+MUX_TO_BNG_IP=$(cat /opt/config/mux_to_bng_net_ipaddr.txt)
+BNG_MUX_CIDR=$(cat /opt/config/bng_mux_net_cidr.txt)
+BRG_BNG_CIDR=$(cat /opt/config/brg_bng_net_cidr.txt)
+BNG_TO_MUX_IP=$(cat /opt/config/bng_to_mux_net_ipaddr.txt)
# Convert Network CIDR to Netmask
cdr2mask () {
# OpenStack network configuration
if [[ $CLOUD_ENV == "openstack" ]]
then
- echo 127.0.0.1 $(hostname) >> /etc/hosts
-
- # Allow remote login as root
- mv /root/.ssh/authorized_keys /root/.ssh/authorized_keys.bk
- cp /home/ubuntu/.ssh/authorized_keys /root/.ssh
-
- MTU=$(/sbin/ifconfig | grep MTU | sed 's/.*MTU://' | sed 's/ .*//' | sort -n | head -1)
-
- IP=$(cat /opt/config/bng_mux_net_ipaddr.txt)
- BITS=$(cat /opt/config/bng_mux_net_cidr.txt | cut -d"/" -f2)
- NETMASK=$(cdr2mask $BITS)
- echo "auto eth1" >> /etc/network/interfaces
- echo "iface eth1 inet static" >> /etc/network/interfaces
- echo " address $IP" >> /etc/network/interfaces
- echo " netmask $NETMASK" >> /etc/network/interfaces
- echo " mtu $MTU" >> /etc/network/interfaces
-
- IP=$(cat /opt/config/oam_ipaddr.txt)
- BITS=$(cat /opt/config/oam_cidr.txt | cut -d"/" -f2)
- NETMASK=$(cdr2mask $BITS)
- echo "auto eth2" >> /etc/network/interfaces
- echo "iface eth2 inet static" >> /etc/network/interfaces
- echo " address $IP" >> /etc/network/interfaces
- echo " netmask $NETMASK" >> /etc/network/interfaces
- echo " mtu $MTU" >> /etc/network/interfaces
-
- IP=$(cat /opt/config/mux_gw_net_ipaddr.txt)
- BITS=$(cat /opt/config/mux_gw_net_cidr.txt | cut -d"/" -f2)
- NETMASK=$(cdr2mask $BITS)
- echo "auto eth3" >> /etc/network/interfaces
- echo "iface eth3 inet static" >> /etc/network/interfaces
- echo " address $IP" >> /etc/network/interfaces
- echo " netmask $NETMASK" >> /etc/network/interfaces
- echo " mtu $MTU" >> /etc/network/interfaces
-
- ifup eth1
- ifup eth2
- ifup eth3
+ echo 127.0.0.1 $(hostname) >> /etc/hosts
+
+ # Allow remote login as root
+ mv /root/.ssh/authorized_keys /root/.ssh/authorized_keys.bk
+ cp /home/ubuntu/.ssh/authorized_keys /root/.ssh
+
+ MTU=$(/sbin/ifconfig | grep MTU | sed 's/.*MTU://' | sed 's/ .*//' | sort -n | head -1)
+
+ IP=$(cat /opt/config/oam_ipaddr.txt)
+ BITS=$(cat /opt/config/oam_cidr.txt | cut -d"/" -f2)
+ NETMASK=$(cdr2mask $BITS)
+ echo "auto eth2" >> /etc/network/interfaces
+ echo "iface eth2 inet static" >> /etc/network/interfaces
+ echo " address $IP" >> /etc/network/interfaces
+ echo " netmask $NETMASK" >> /etc/network/interfaces
+ echo " mtu $MTU" >> /etc/network/interfaces
+
+ ifup eth2
fi
-# Download required dependencies
-add-apt-repository -y ppa:openjdk-r/ppa
-apt-get update
-apt-get install -y wget openjdk-8-jdk apt-transport-https ca-certificates g++ libcurl4-gnutls-dev
-sleep 1
+# Auto-start configuration for the VPP
+cat > /etc/vpp/startup.conf << EOF
+
+unix {
+ nodaemon
+ log /tmp/vpp.log
+ full-coredump
+ cli-listen localhost:5002
+ startup-config /etc/vpp/setup.gate
+}
+
+api-trace {
+ on
+}
+
+api-segment {
+ gid vpp
+}
+
+cpu {
+ ## In the VPP there is one main thread and optionally the user can create worker(s)
+ ## The main thread and worker thread(s) can be pinned to CPU core(s) manually or automatically
+
+ ## Manual pinning of thread(s) to CPU core(s)
+
+ ## Set logical CPU core where main thread runs
+ # main-core 1
+
+ ## Set logical CPU core(s) where worker threads are running
+ # corelist-workers 2-3,18-19
+
+ ## Automatic pinning of thread(s) to CPU core(s)
+
+ ## Sets number of CPU core(s) to be skipped (1 ... N-1)
+ ## Skipped CPU core(s) are not used for pinning main thread and working thread(s).
+ ## The main thread is automatically pinned to the first available CPU core and worker(s)
+ ## are pinned to next free CPU core(s) after core assigned to main thread
+ # skip-cores 4
+
+ ## Specify a number of workers to be created
+ ## Workers are pinned to N consecutive CPU cores while skipping "skip-cores" CPU core(s)
+ ## and main thread's CPU core
+ # workers 2
+
+ ## Set scheduling policy and priority of main and worker threads
+
+ ## Scheduling policy options are: other (SCHED_OTHER), batch (SCHED_BATCH)
+ ## idle (SCHED_IDLE), fifo (SCHED_FIFO), rr (SCHED_RR)
+ # scheduler-policy fifo
+
+ ## Scheduling priority is used only for "real-time policies (fifo and rr),
+ ## and has to be in the range of priorities supported for a particular policy
+ # scheduler-priority 50
+}
+
+# dpdk {
+ ## Change default settings for all intefaces
+ # dev default {
+ ## Number of receive queues, enables RSS
+ ## Default is 1
+ # num-rx-queues 3
+
+ ## Number of transmit queues, Default is equal
+ ## to number of worker threads or 1 if no workers treads
+ # num-tx-queues 3
+
+ ## Number of descriptors in transmit and receive rings
+ ## increasing or reducing number can impact performance
+ ## Default is 1024 for both rx and tx
+ # num-rx-desc 512
+ # num-tx-desc 512
+
+ ## VLAN strip offload mode for interface
+ ## Default is off
+ # vlan-strip-offload on
+ # }
+
+ ## Whitelist specific interface by specifying PCI address
+ # dev 0000:02:00.0
+
+ ## Whitelist specific interface by specifying PCI address and in
+ ## addition specify custom parameters for this interface
+ # dev 0000:02:00.1 {
+ # num-rx-queues 2
+ # }
+
+ ## Change UIO driver used by VPP, Options are: igb_uio, vfio-pci
+ ## and uio_pci_generic (default)
+ # uio-driver vfio-pci
+
+ ## Disable mutli-segment buffers, improves performance but
+ ## disables Jumbo MTU support
+ # no-multi-seg
+
+ ## Increase number of buffers allocated, needed only in scenarios with
+ ## large number of interfaces and worker threads. Value is per CPU socket.
+ ## Default is 16384
+ # num-mbufs 128000
+
+ ## Change hugepages allocation per-socket, needed only if there is need for
+ ## larger number of mbufs. Default is 256M on each detected CPU socket
+ # socket-mem 2048,2048
+# }
+
+EOF
+
+# Get list of network device PCI bus addresses
+get_nic_pci_list() {
+ while read -r line ; do
+ if [ "$line" != "${line#*network device}" ]; then
+ echo -n "${line%% *} "
+ fi
+ done < <(lspci)
+}
+
+NICS=$(get_nic_pci_list)
+NICS=`echo ${NICS} | sed 's/[0]\+\([0-9]\)/\1/g' | sed 's/[.:]/\//g'`
+
+BNG_MUX_NIC=GigabitEthernet`echo ${NICS} | cut -d " " -f 2` # second interface in list
+MUX_GW_NIC=GigabitEthernet`echo ${NICS} | cut -d " " -f 4` # fourth interface in list
+
+cat > /etc/vpp/setup.gate << EOF
+set int state ${BNG_MUX_NIC} up
+set int ip address ${BNG_MUX_NIC} ${MUX_TO_BNG_IP}/${BNG_MUX_CIDR#*/}
+
+set int state ${MUX_GW_NIC} up
+set int ip address ${MUX_GW_NIC} ${MUX_GW_IP}/${MUX_GW_CIDR#*/}
+ip route add ${BRG_BNG_CIDR} via ${BNG_TO_MUX_IP} ${BNG_MUX_NIC}
+EOF
+
+
+# Create systemctl service for Honeycomb
+cat > /etc/systemd/system/honeycomb.service << EOF
+[Unit]
+Description=Honeycomb Agent for the VPP control plane
+Documentation=https://wiki.fd.io/view/Honeycomb
+Requires=vpp.service
+After=vpp.service
+
+[Service]
+ExecStart=/opt/honeycomb/honeycomb
+ExecStop=/opt/reset_ves_mode.sh
+Restart=always
+RestartSec=10
+
+[Install]
+WantedBy=multi-user.target
+EOF
+systemctl enable /etc/systemd/system/honeycomb.service
+
+# Create script that is used to reset the ves mode configuration
+# when the honeycomb service is stopped
+cat > /opt/reset_ves_mode.sh << EOF
+#!/bin/bash
+mode_data=\$(curl -X GET -u admin:admin http://127.0.0.1:8183/restconf/config/vesagent:vesagent/mode 2>/dev/null | sed s/\"base-packet-loss\":[0-9]\*,/\"base-packet-loss\":0,/)
+if [[ "\${mode_data}" != "" ]]
+then
+ curl -X DELETE -u admin:admin http://127.0.0.1:8183/restconf/config/vesagent:vesagent/mode
+ curl -H "Content-Type:application/json" --data "\${mode_data}" -X POST -u admin:admin http://127.0.0.1:8183/restconf/config/vesagent:vesagent
+fi
+EOF
+ chmod a+x /opt/reset_ves_mode.sh
+
+ #Create a systemd service for auto-save
+ cat > /usr/bin/save_config << EOF
+#!/bin/bash
+
+#########################################################################
+#
+# Copyright (c) 2017 Intel and/or its affiliates.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at:
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#
+##########################################################################
+
+############################### Variables ################################
+VPP_SETUP_GATE=/etc/vpp/setup.gate
+
+############################### Functions ################################
+
+# Write the commands to the startup scripts.
+#
+# We could set VPP configuration to the startup.conf.
+# Write the configuration to the startup scripts so we could
+# restore the system after rebooting.
+#
+write_startup_scripts()
+{
+ local cmd=\${2}
+ local is_add=\${1}
+
+ if [[ \${is_add} == add ]] ;then
+ while read -r line
+ do
+ if [[ \${line} == \${cmd} ]] ;then
+ return 0
+ fi
+ done < \${VPP_SETUP_GATE}
+
+ echo "\${cmd}" >> \${VPP_SETUP_GATE}
+ else
+ while read -r line
+ do
+ if [[ \${line} == \${cmd} ]] ;then
+ sed -i "/\${line}/d" \${VPP_SETUP_GATE}
+ return 0
+ fi
+ done < \${VPP_SETUP_GATE}
+ fi
+}
+
+# Save the VxLAN Tunnel Configuration to the startup script.
+#
+# Get the current VxLAN tunnel configuration with bash command:
+# \$vppctl show vxlan tunnel
+# [0] src 10.3.0.2 dst 10.1.0.20 vni 100 sw_if_index 1 encap_fib_index 0 fib_entry_index 7 decap_next l2
+# [1] src 10.5.0.20 dst 10.5.0.21 vni 100 sw_if_index 2 encap_fib_index 0 fib_entry_index 8 decap_next l2
+# Set the VxLAN Tunnel with the bash command:
+# \$vppctl create vxlan tunnel src 10.3.0.2 dst 10.1.0.20 vni 100
+# vxlan_tunnel0
+save_vxlan_tunnel()
+{
+ local src=""
+ local dst=""
+ local vni=""
+
+ vppctl show vxlan tunnel | while read line
+ do
+ if [ "\${line}" != "" ] ;then
+ src=\`echo \${line} | awk '{ print \$3 }'\`
+ dst=\`echo \${line} | awk '{ print \$5 }'\`
+ vni=\`echo \${line} | awk '{ print \$7 }'\`
+
+ write_startup_scripts add "create vxlan tunnel src \${src} dst \${dst} vni \${vni}"
+ fi
+ done
+}
+
+# Save the VxLAN tunnel L2 xconnect configuration to the startup script.
+#
+# Get the Current L2 Address configuration with bash command:
+# \$vppctl show int addr
+# local0 (dn):
+# vxlan_tunnel0 (up):
+# l2 xconnect vxlan_tunnel1
+# vxlan_tunnel1 (up):
+# l2 xconnect vxlan_tunnel0
+# Save the VxLAN tunnel L2 xconnect configuration with bash command:
+# \$vppctl set interface l2 xconnect vxlan_tunnel0 vxlan_tunnel1
+#
+save_vxlan_xconnect()
+{
+ local ingress=""
+ local egress=""
+
+ vppctl show int addr | while read line
+ do
+ if [[ \${line} == vxlan_tunnel* ]] ;then
+ read next
+ while [[ \${next} != l2* ]] || [[ \${next} == "" ]]
+ do
+ line=\`echo \${next}\`
+ read next
+ done
+ if [[ \${next} == l2* ]] ;then
+ ingress=\`echo \${line} | awk '{ print \$1 }'\`
+ egress=\`echo \${next} | awk '{ print \$3 }'\`
+ write_startup_scripts add "set interface l2 xconnect \${ingress} \${egress}"
+ fi
+ fi
+ done
+}
+
+################################# MAIN ###################################
+
+save_vxlan_tunnel
+
+save_vxlan_xconnect
+
+EOF
+ chmod a+x /usr/bin/save_config
+ cat > /etc/systemd/system/autosave.service << EOF
+[Unit]
+Description=Run Scripts at Start and Stop
+equires=vpp.service
+After=vpp.service
+
+[Service]
+Type=oneshot
+RemainAfterExit=true
+ExecStop=/usr/bin/save_config
+
+[Install]
+WantedBy=multi-user.target
+EOF
+systemctl enable /etc/systemd/system/autosave.service
# Download DHCP config files
cd /opt
-wget $REPO_URL_BLOB/org.onap.demo/vnfs/vcpe/$INSTALL_SCRIPT_VERSION/v_gmux_init.sh
-wget $REPO_URL_BLOB/org.onap.demo/vnfs/vcpe/$INSTALL_SCRIPT_VERSION/v_gmux.sh
+unzip -p -j /opt/vcpe-scripts-$SCRIPT_VERSION.zip v_gmux_init.sh > /opt/v_gmux_init.sh
+unzip -p -j /opt/vcpe-scripts-$SCRIPT_VERSION.zip v_gmux.sh > /opt/v_gmux.sh
chmod +x v_gmux_init.sh
chmod +x v_gmux.sh
mv v_gmux.sh /etc/init.d
# Rename network interface in openstack Ubuntu 16.04 images. Then, reboot the VM to pick up changes
if [[ $CLOUD_ENV != "rackspace" ]]
then
- sed -i "s/GRUB_CMDLINE_LINUX=.*/GRUB_CMDLINE_LINUX=\"net.ifnames=0 biosdevname=0\"/g" /etc/default/grub
- grub-mkconfig -o /boot/grub/grub.cfg
- sed -i "s/ens[0-9]*/eth0/g" /etc/network/interfaces.d/*.cfg
- sed -i "s/ens[0-9]*/eth0/g" /etc/udev/rules.d/70-persistent-net.rules
- echo 'network: {config: disabled}' >> /etc/cloud/cloud.cfg.d/99-disable-network-config.cfg
- echo "APT::Periodic::Unattended-Upgrade \"0\";" >> /etc/apt/apt.conf.d/10periodic
- reboot
+ sed -i "s/GRUB_CMDLINE_LINUX=.*/GRUB_CMDLINE_LINUX=\"net.ifnames=0 biosdevname=0\"/g" /etc/default/grub
+ grub-mkconfig -o /boot/grub/grub.cfg
+ sed -i "s/ens[0-9]*/eth0/g" /etc/network/interfaces.d/*.cfg
+ sed -i "s/ens[0-9]*/eth0/g" /etc/udev/rules.d/70-persistent-net.rules
+ echo 'network: {config: disabled}' >> /etc/cloud/cloud.cfg.d/99-disable-network-config.cfg
+ echo 'Execution of vG-MUX install script completed' > /opt/script_status.txt
+ reboot
fi
-# Run docker containers. For openstack Ubuntu 16.04 images this will run as a service after the VM has restarted
-./v_gmux_init.sh
\ No newline at end of file
+./v_gmux_init.sh
+echo 'Execution of vG-MUX install script completed' > /opt/script_status.txt
+