-
Notifications
You must be signed in to change notification settings - Fork 2
/
enabler_conf.ini
225 lines (188 loc) · 6.73 KB
/
enabler_conf.ini
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
[general]
# Compute node user name and password that can be used to ssh command.
# Also the user must be a sudoer assuming all compute nodes have the same
# password.
# Credentials of compute servers.
#
# compute_user = username
# compute_passwd = password
# IF uplink detection is not needed, enter the hostname in node field, which
# is comma separated list of hosts for which static uplink is configured.
# The node_uplink field has the comma separated uplink ports in server which
# is connected to the leaf. In example below, host1 has eth5 connected to
# leaf and host2 has eth4 connected to leaf.
# The node name should be fully qualified domain name (e.g. host1.example.com)
#
# node = host1.example.com, host2.example.com
# node_uplink = eth5, eth4
#
# If Openstack is running in any UCS FI blade server, enter the EVB DMAC
# configured in the fabric. The enabler software running in the node will
# detect if it's a UCS FI blade server, but the interface connected to the
# switch has be given in the 'node_uplink' configuration above along with
# the node.
#
# ucs_fi_evb_dmac = '01:80:c2:12:34:56'
[dfa_agent]
# OVS Neutron Agent related configuration. Ensure that this is the same as
# what is configured for OVS Neutron Agent.
# The defaults are given below for convenience.
# integration_bridge = br-int
# external_dfa_bridge = br-ethd
[dcnm]
# IP address of the DCNM. It should be reachable from openstack
# controller node.
#
# dcnm_ip = 3.3.3.3
# DCNM server login and rabbitmq messaging credentials
#
# dcnm_user = dcnm-login-username
# dcnm_amqp_user = dncm-amqp-username
# dcnm_password = dcnm-password
# Gateway MAC address. It should be the same as configured on the leaf nodes
#
# gateway_mac = 20:20:00:00:00:AA
# Orchestrator ID
# Orchestartor id used for registering the segmentation id range on DCNM
# If there are multiple setups using the same DCNM, please ensure
# different orchestrator IDs are used.
#
# orchestrator_id = 'Openstack Controller'
# Segmentation ID range.
# The seg-id-min and seg-id-max are 24-bit integer values
#
# segmentation_id_min = seg-id-min
# segmentation_id_max = seg-id-max
# This means the segmentation ID after released will only be reused after
# 1 hour. If this functionality is not needed, enter a value of 0. OR if you
# would like to change the default of 1hr, uncomment the below and enter a
# different number (only integers)
# segmentation_reuse_timeout = 1
# Currently this is needed only for services like LB and FW.
# If the below is left commented, the defaults will be picked, which is also
# shown below for convenience.
# Please ensure the range falls in the detectable range of the mobility
# domain of fabric leaf ports where FW or LB are connected.
#vlan_id_min = 800
#vlan_id_max = 1000
# Same meaning as that of seg reuse above. But, here the default is 0, which
# means the VLAN will be reused immediately after release.
#vlan_reuse_timeout = 0
# The suffix of a network name when it is created by DCNM.
#
# dcnm_net_ext = '(DCNM)'
# The lease file name of DHCP server on the DCNM.
#
# dcnm_dhcp_leases = '/var/lib/dhcpd/dhcpd.leases'
# Default configuration profile when creating a network in DCNM.
#
# The default is 'defaultNetworkUniversalTfProfile'.
# default_cfg_profile = defaultNetworkUniversalTfProfile
# Default vrf profile name for a partition in DCNM.
#
# default_vrf_profile = vrf-common-universal
# If you want to use dcnm's dhcp server, then set dcnm_dhcp=True,
# If you set it to false, openstack native dhcp will be used instead
# dcnm_dhcp = True
[dfa_rpc]
# Transport URL parameter for RPC.
# The credentials should be set based on setup.
#
# transport_url = rabbit://username:password@rabbitmq-server-ip:5672//
[dfa_mysql]
# MYSQL DB connection option
# The credentials should be set based on the setup.
#
# connection = mysql://username:password@mysql-server-ip/cisco_dfa?charset=utf8
[dfa_notify]
# Notification queue name for DFA enabler.
# service_name: keystone and neutron
#
# cisco_dfa_notify_queue = cisco_dfa_%(service_name)s_notify
[dfa_log]
# Log file name. DEPRECATED (use Log file prefix instead)
# If log file name and directory is not specified, the default
# is the standard output.
#
# log_file = fabric_enabler.log
# Log file prefix
# Prefix will be used by Fabric Enabler processes to
# create own log files.
#
# log_file_prefix = fabric_enabler
# The direcoty name for the log file.
#
# log_dir = /home/localadmin/Logs
# Enabler debugging output level. Default is WARNING.
# Set to DEBUG to see the debbugging output
#
# log_level = DEBUG
# Enable syslog. Debugging messages goes to syslog
#
# use_syslog = False
# syslog_log_facility = LOG_USER
# Output format of log messages.
#
# log_format = %(asctime)s %(levelname)8s [%(name)s] %(message)s
# If need to save pid of fabric-enabler-server and fabric-enabler-agent
# the location of pid files should be set by the following options.
#
# pid_dir =
# pid_server_file =
# pid_agent_file =
[vdp]
# VDP default options
#
# mgrid2 = 0
# typeid = 0
# typeidver = 0
# vsiidfrmt = 5
# hints = none
# filter = 4
# vdp_sync_timeout = 30
[firewall]
# Firewall Default Parameters
#
#Currently supports phy_asa and Openstack native Firewall.
#Uncomment the below if FW support is needed and enter phy_asa or native
#device = phy_asa
#fw_mgmt_ip = [1.1.1.1]
#Uncomment the below and fill up the right values for phy_asa, not needed for
#native FW.
#fw_username = [admin]
#fw_password = [cisco123]
#fw_interface_in = [Gi0/0]
#fw_interface_out = [Gi0/1]
# Currently only the MAX scheduling is supported
#sched_policy = 'max_sched'
# fw_auto_serv_nwk_create = True
# mob_domain_name = 'md0'
# fw_service_host_profile = 'serviceNetworkUniversalESChainLBESProfile'
# fw_service_host_fwd_mode = 'proxy-gateway'
# fw_service_part_vrf_profile = 'vrf-common-universal-external-dynamic-ES'
# fw_service_ext_profile = 'externalNetworkUniversalDynamicRoutingESProfile'
# fw_service_ext_fwd_mode = 'anycast-gateway'
# It assumes a 24 subnet. In the example below, the first service IN network
# will have 100.100.2.x, and subsequent ones withh have 100.100.3.x and so on.
# fw_service_in_ip_start = 100.100.2.0
# fw_service_in_ip_end = 100.100.20.0
# fw_service_out_ip_start = 200.200.2.0
# fw_service_out_ip_end = 200.200.20.0
# fw_service_dummy_ip_subnet = '9.9.9.0/24'
[loadbalance]
#lb_enabled = false
#lb_vrf_profile = vrf-common-universal-dynamic-LB-ES
#lb_svc_net_profile=serviceNetworkUniversalDynamicRoutingLBProfile
#lb_svc_net = 199.199.1.0/24
#lb_svc_net_name_prefix=lbaasInternal
#lb_driver = dfa.server.services.loadbalance.drivers.f5.F5Device.F5Device
# The following must be filled by the user
# comma seperated list of mgmt IPs of the box
#lb_mgmt_ip = 172.28.10.180
#lb_user_name = admin
#lb_user_password = cisco123
#lb_f5_interface = 1.1
[sys]
# Default root_helper
#
# root_helper = 'sudo'