ON THIS PAGE
Download jinstall-vmx-14.1R1.10-domestic.img at FILENINJA.NET. This file (jinstall-vmx-14.1R1.10-domestic.img) is hosted at free file sharing service 4shared. If you are the copyright owner for this file, please Report Abuse to 4shared.
Read this topic to understand how to installvMX instance in the OpenStack environment.
Preparing the OpenStack Environment to Install vMX
Make sure the openstackrc
file is sourced before you run any OpenStack commands.
To prepare the OpenStack environment to install vMX, performthese tasks:
Creating the neutron Networks
You must create the neutron networks used by vMX before youstart the vMX instance. The public network is the neutron networkused for the management (fxp0) network. The WAN network is the neutronnetwork on which the WAN interface forvMX is added.
To display the neutron network names, use the network-name --shared--provider:physical_network network-namesubnetwork-name --allocation-poolstart=end-address --gateway=network-name --router:external=True--provider:network_type vlan --provider:physical_network segment-id
neutron subnet-create address --name start-address,end=gateway-address
For example:
neutron net-create OSP_PROVIDER_1500 --router:external=True--provider:network_type vlan --provider:physical_network physnet1--provider:segmentation_id 1500
neutron subnet-create OSP_PROVIDER_1500 11.0.2.0/24 --nameOSP_PROVIDER_1500_SUBNET --enable_dhcp=False --allocation-pool start=11.0.2.10,end=11.0.2.100--gateway=11.0.2.254
For SR-IOV, you can use these commands as one way to createthe WAN network:
neutron net-create network-name
neutron subnet-create address --name start-address,end=gateway-address
For example:
neutron net-create OSP_PROVIDER_SRIOV --router:external=True--provider:network_type vlan --provider:physical_network physnet2
neutron subnet-create OSP_PROVIDER_SRIOV 12.0.2.0/24 --nameOSP_PROVIDER_SRIOV_SUBNET --enable_dhcp=False --allocation-pool start=12.0.2.10,end=12.0.2.100--gateway=12.0.2.254
Preparing the Controller Node
Preparing the Controller Node for vMX
- Configure the controller node to enable Huge Pages andCPU affinity by editing the systemctl restart openstack-nova-scheduler.service
For Ubuntu (starting with Junos OS Release 17.2R1): Note
We recommend these default values, but you can use differentvalues if they are appropriate for your environment. Make sure thedefault quotas have enough allocated resources.
Verify the changes with the rpm -qa grep heat command.
For Red Hat: apt-get install heat-engine
- Make sure the lsb (redhat-lsb-core or lsb-release) andnumactl packages are installed.
For Red Hat:
For Ubuntu (starting with Junos OS Release 17.2R1):
Configuring the Controller Node for virtio Interfaces
To configure the virtio interfaces:
- Enable the VLAN mechanism driver by adding type_drivers parameter in the
/etc/neutron/plugins/ml2/ml2_conf.ini
file. - Add the bridge mapping to the
/etc/neutron/plugins/ml2/ml2_conf.ini
file by adding the following line:For example, use the following setting to add a bridge mappingfor the physical network physnet1 mapped to the OVS bridge br-vlan.
- Configure the VLAN ranges used for the physical networkin the
/etc/neutron/plugins/ml2/ml2_conf.ini
file, where physical-network-name is the name of the neutron network that you created for the virtioWAN network.For example, use the following setting to configure the VLANranges used for the physical network physnet1.
- Restart the neutron server.
For Red Hat: service neutron-server restart
- Add the OVS bridge that was mapped to the physical networkand the virtio interface (eth2).
For example, use the following commands to add OVS bridge br-vlanand eth2 interface:
Configuring the Controller Node for SR-IOV Interfaces
NoteIf you have more than one SR-IOV interface, you need onephysical 10G Ethernet NIC card for each additional SR-IOV interface.
NoteIn SRIOV mode, the communication between the Routing Engine(RE) and packet forwarding engine is enabled using virtio interfaceson a VLAN-provider OVS network. Because of this, a given physicalinterface cannot be part of both VirtIO and SR-IOV networks.
To configure the SR-IOV interfaces:
- Edit the
/etc/neutron/plugins/ml2/ml2_conf.ini
file to add –-config-file /etc/neutron/plugins/ml2/ml2_conf_sriov.ini as highlighted to the neutron server file.For Red Hat:
Edit the
/usr/lib/systemd/system/neutron-server.service
file as highlighted.Use the service neutron-server restart command torestart the service.
- To allow proper scheduling of SR-IOV devices, the computescheduler must use the FilterScheduler with the PciPassthroughFilterfilter.
Make sure the PciPassthroughFilter filter is configured in the
/etc/nova/nova.conf
file on the controller node.For Red Hat: service nova-scheduler restart
Preparing the Compute Nodes
Preparing the Compute Node for vMX
NoteYou no longer need to configure the compute node to passmetadata to the vMX instances by including the
For Red Hat: Add the Huge Pages configuration.
Use the GRUB_CMDLINE_LINUX_DEFAULT parameter.
After the reboot, verify that Huge Pages are allocated.
The number of Huge Pages depends on the amount of memory forthe VFP, the size of Huge Pages, and the number of VFP instances.To calculate the number of Huge Pages: (memory-for-vfp / huge-pages-size) * number-of-vfp
For example, if you run four vMX instances (four VFPs) in performancemode using 12G of memory and 2M of Huge Pages size, then the numberof Huge Pages as calculated by the formula is (12G/2M)*4 or 24576.
Starting in Junos OS Release 15.1F6 and in later releases,performance mode is the default operating mode. For details, see Enabling Performance Mode or Lite Mode.
Ensure that you have enough physical memory on the computenode. It must be greater than the amount of memory allocated to HugePages because any other applications that do not use Huge Pages arelimited by the amount of memory remaining after allocation for HugePages. For example, if you allocate 24576 Huge Pages and 2M Huge Pagessize, you need 24576*2M or 48G of memory for Huge Pages.
You can use the intel_iommu=on string to any existingtext for the grub2-mkconfig -o /boot/grub2/grub.cfg
For Ubuntu (starting with Junos OS Release 17.2R1): br-vlan is added. (Thisis the same physnet1:br-vlan string:
Restart neutron service.
Redhat:
systemctl restart openstack-nova-compute.service
Ubuntu
service neutron-plugin-openvswitch-agent restart
Configuring the Compute Node for SR-IOV Interfaces
NoteIf you have more than one SR-IOV interface, you need onephysical 10G Ethernet NIC card for each additional SR-IOV interface.
- Load the modified IXGBE driver.
Before compiling the driver, make sure gcc and make are installed.
For Red Hat:
For Ubuntu (starting with Junos OS Release 17.2R1):
Unload the default IXGBE driver, compile the modified JuniperNetworks driver, and load the modified IXGBE driver.
Verify the driver version on the eth4 interface.
For example, in the following sample, the command displays driverversion (3.19.1):
- Create the virtual function (VF) on the physical device.vMX currently supports only one VF for each SR-IOV interface (forexample, eth4).
Specify the number of VFs on each NIC. The following line specifiesthat there is no VF for eth2 (first NIC) and one VF for eth4 (secondNIC with SR-IOV interface).
To verify that the VF was created, the output of the sudo yum install openstack-neutron-sriov-nic-agent
For Ubuntu (starting with Junos OS Release 17.2R1): –-config-file/etc/neutron/plugins/ml2/sriov_agent.ini as highlighted.
For Red Hat:
Edit the
/usr/lib/systemd/system/neutron-sriov-nic-agent.service
file as highlighted.Enable and start the SR-IOV agent.
Use the service neutron-plugin-sriov-agent start commandto start the SR-IOV agent.
Use the systemctl restart openstack-nova-compute
For Ubuntu (starting with Junos OS Release 17.2R1):
Setting Up the vMX Configuration File
The parameters required to configure vMX are defined in thestartup configuration file.
- Download the vMX KVM software package from the vMX page and uncompress the package.
package-name
- Change directory to the location of the files.
package-location/openstack/scripts
- Edit the
vmx.conf
textfile with a text editor to create the flavors for a single vMX instance.Based on your requirements, ensure the following parametersare set properly in the vMX configuration file:
pfe-flavor-name
memory-mb
See Specifying vMX Configuration File Parameters for informationabout the parameters.
Sample vMX Startup Configuration File
Here is a sample vMX startup configuration file for OpenStack:
See also
Specifying vMX Configuration File Parameters
The parameters required to configure vMX aredefined in the startup configuration file (
scripts/vmx.conf
). The startup configuration file generates a file that is used tocreate flavors. To create new flavors with different memory-mb parameters, you must change the corresponding pfe-flavor-name parameter beforecreating the new flavors.To customize the configuration, perform these tasks:
Configuring the Host
To configure the host, navigate to virtualization-type
- Download the vMX KVM software package from the vMX page and uncompress the package.
If this parameter is not specified, the output of the nova hypervisor-listcommand provides the list of compute nodes on which to run vMX instances.
Configuring the VCP VM
To configure the VCP VM, you must provide the flavor name.
NoteWe recommend unique values for the CONTROL_PLANE and specify the following parameters:
vcpus—Number of vCPUs for the VCP; minimumis 1.
Note If you change this value, you must change the memory-mb—Amount of memory for the VCP;minimum is 4 GB.
Note If you change this value, you must change the FORWARDING_PLANE and specify the following parameters:
memory-mb—Amount of memory for the VFP;minimum is 12 GB (performance mode) and 4 GB (lite mode).
Note If you change this value, you must change the vcpus—Number of vCPUs for the VFP; minimumis 7 (performance mode) and 3 (lite mode).
Note If you specify less than 7 vCPUs, the VFP automaticallyswitches to lite mode.
Note If you change this value, you must change the
- Run the vmx_osp_flavors.sh filethat creates flavors.
vmx_osp_flavors.sh to create flavors.
Installing vMX Images for the VCP and VFP
To install the vMX OpenStack glance images for the VCP and VFP,you can execute the
- Download the vMX KVM software package from the vMX page and uncompress the package.
package-name
- Verify the location of the software images from the uncompressedvMX package. See vMX Package Contents.
package-location/images
- Change directory to the location of the vMX OpenStackscript files.
package-location/openstack/scripts
- Run the sh vmx_osp_images.sh vcp-image-locationvfp-image-location
Note You must specify the parameters in this order.
vcp-image-location—Absolute path to the
junos-vmx-x86-64*.qcow2
file for launching VCP.vfp-image-location—Absolute path to the
vFPC-*.img
file for launching VFP.
Modify these parameters in the
net_id1
—Network ID of the existing neutronnetwork used for the WAN port. Use the public_network—Network ID of the existingneutron network used for the management (fxp0) port. Use the fpc_img—Change this parameter to vfp-image-name parameter specified when running the scriptto install the vMX images.vfp-image-name parameter specifiedwhen running the script to install the vMX images (applicable forFor Junos OS Releases 17.3R1 and earlier).
linux-flav. Name of the nova flavor for the VFP; same as the vfp_flavor—Name of the nova flavor forthe VFP; same as the junos_flav—Name of the nova flavor forthe VCP; same as the vcp_flavor—Name of the nova flavor forthe VCP; same as the junos_img—Name of the glance image forthe VCP; same as the vcp_image—Name of the glance image forthe VCP; same as the project_name—Any project name. All resourceswill use this name as the prefix.
heat stack-create–f 1vmx.yaml –e 1vmx.env heat stack-list grep nova-list command.
- Access the VCP or the VFP VM with the nova-id novnc command, where nova-list commandoutput.Note
For example, this command installs the VCP image as re-testfrom the
/var/tmp/junos-vmx-x86-64-17.1R1.8.qcow2
file and the VFP image as fpc-test from the/var/tmp/vFPC-20170117.img
file.glance image-list command.
Starting a vMX Instance
Modifying Initial Junos OS Configuration
When you start the vMX instance, the Junos OS configurationfile found in
vmx_baseline.conf file or move the file, make sure that the
You must shut down the vMX instance before you reboothost server using the request system halt command.
Related Documentation