1 /*
2 * Copyright (c) 2017-2018 Cavium, Inc.
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 *
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
16 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25 * POSSIBILITY OF SUCH DAMAGE.
26 *
27 * $FreeBSD$
28 *
29 */
30
31 #ifndef __ECORE_SRIOV_API_H__
32 #define __ECORE_SRIOV_API_H__
33
34 #include "common_hsi.h"
35 #include "ecore_status.h"
36
37 #define ECORE_ETH_VF_NUM_MAC_FILTERS 1
38 #define ECORE_ETH_VF_NUM_VLAN_FILTERS 2
39 #define ECORE_VF_ARRAY_LENGTH (3)
40
41 #define IS_VF(p_dev) ((p_dev)->b_is_vf)
42 #define IS_PF(p_dev) (!((p_dev)->b_is_vf))
43 #ifdef CONFIG_ECORE_SRIOV
44 #define IS_PF_SRIOV(p_hwfn) (!!((p_hwfn)->p_dev->p_iov_info))
45 #else
46 #define IS_PF_SRIOV(p_hwfn) (0)
47 #endif
48 #define IS_PF_SRIOV_ALLOC(p_hwfn) (!!((p_hwfn)->pf_iov_info))
49 #define IS_PF_PDA(p_hwfn) 0 /* @@TBD Michalk */
50
51 /* @@@ TBD MichalK - what should this number be*/
52 #define ECORE_MAX_VF_CHAINS_PER_PF 16
53
54 /* vport update extended feature tlvs flags */
55 enum ecore_iov_vport_update_flag {
56 ECORE_IOV_VP_UPDATE_ACTIVATE = 0,
57 ECORE_IOV_VP_UPDATE_VLAN_STRIP = 1,
58 ECORE_IOV_VP_UPDATE_TX_SWITCH = 2,
59 ECORE_IOV_VP_UPDATE_MCAST = 3,
60 ECORE_IOV_VP_UPDATE_ACCEPT_PARAM = 4,
61 ECORE_IOV_VP_UPDATE_RSS = 5,
62 ECORE_IOV_VP_UPDATE_ACCEPT_ANY_VLAN = 6,
63 ECORE_IOV_VP_UPDATE_SGE_TPA = 7,
64 ECORE_IOV_VP_UPDATE_MAX = 8,
65 };
66
67 /*PF to VF STATUS is part of vfpf-channel API
68 and must be forward compatible */
69 enum ecore_iov_pf_to_vf_status {
70 PFVF_STATUS_WAITING = 0,
71 PFVF_STATUS_SUCCESS,
72 PFVF_STATUS_FAILURE,
73 PFVF_STATUS_NOT_SUPPORTED,
74 PFVF_STATUS_NO_RESOURCE,
75 PFVF_STATUS_FORCED,
76 PFVF_STATUS_MALICIOUS,
77 };
78
79 struct ecore_mcp_link_params;
80 struct ecore_mcp_link_state;
81 struct ecore_mcp_link_capabilities;
82
83 /* These defines are used by the hw-channel; should never change order */
84 #define VFPF_ACQUIRE_OS_LINUX (0)
85 #define VFPF_ACQUIRE_OS_WINDOWS (1)
86 #define VFPF_ACQUIRE_OS_ESX (2)
87 #define VFPF_ACQUIRE_OS_SOLARIS (3)
88 #define VFPF_ACQUIRE_OS_LINUX_USERSPACE (4)
89 #define VFPF_ACQUIRE_OS_FREEBSD (5)
90
91 struct ecore_vf_acquire_sw_info {
92 u32 driver_version;
93 u8 os_type;
94 };
95
96 struct ecore_public_vf_info {
97 /* These copies will later be reflected in the bulletin board,
98 * but this copy should be newer.
99 */
100 u8 forced_mac[ETH_ALEN];
101 u16 forced_vlan;
102 };
103
104 struct ecore_iov_vf_init_params {
105 u16 rel_vf_id;
106
107 /* Number of requested Queues; Currently, don't support different
108 * number of Rx/Tx queues.
109 */
110 /* TODO - remove this limitation */
111 u16 num_queues;
112
113 /* Allow the client to choose which qzones to use for Rx/Tx,
114 * and which queue_base to use for Tx queues on a per-queue basis.
115 * Notice values should be relative to the PF resources.
116 */
117 u16 req_rx_queue[ECORE_MAX_VF_CHAINS_PER_PF];
118 u16 req_tx_queue[ECORE_MAX_VF_CHAINS_PER_PF];
119
120 u8 vport_id;
121
122 /* Should be set in case RSS is going to be used for VF */
123 u8 rss_eng_id;
124 };
125
126 #ifdef CONFIG_ECORE_SW_CHANNEL
127 /* This is SW channel related only... */
128 enum mbx_state {
129 VF_PF_UNKNOWN_STATE = 0,
130 VF_PF_WAIT_FOR_START_REQUEST = 1,
131 VF_PF_WAIT_FOR_NEXT_CHUNK_OF_REQUEST = 2,
132 VF_PF_REQUEST_IN_PROCESSING = 3,
133 VF_PF_RESPONSE_READY = 4,
134 };
135
136 struct ecore_iov_sw_mbx {
137 enum mbx_state mbx_state;
138
139 u32 request_size;
140 u32 request_offset;
141
142 u32 response_size;
143 u32 response_offset;
144 };
145
146 /**
147 * @brief Get the vf sw mailbox params
148 *
149 * @param p_hwfn
150 * @param rel_vf_id
151 *
152 * @return struct ecore_iov_sw_mbx*
153 */
154 struct ecore_iov_sw_mbx*
155 ecore_iov_get_vf_sw_mbx(struct ecore_hwfn *p_hwfn,
156 u16 rel_vf_id);
157 #endif
158
159 /* This struct is part of ecore_dev and contains data relevant to all hwfns;
160 * Initialized only if SR-IOV cpabability is exposed in PCIe config space.
161 */
162 struct ecore_hw_sriov_info {
163 /* standard SRIOV capability fields, mostly for debugging */
164 int pos; /* capability position */
165 int nres; /* number of resources */
166 u32 cap; /* SR-IOV Capabilities */
167 u16 ctrl; /* SR-IOV Control */
168 u16 total_vfs; /* total VFs associated with the PF */
169 u16 num_vfs; /* number of vfs that have been started */
170 u16 initial_vfs; /* initial VFs associated with the PF */
171 u16 nr_virtfn; /* number of VFs available */
172 u16 offset; /* first VF Routing ID offset */
173 u16 stride; /* following VF stride */
174 u16 vf_device_id; /* VF device id */
175 u32 pgsz; /* page size for BAR alignment */
176 u8 link; /* Function Dependency Link */
177
178 u32 first_vf_in_pf;
179 };
180
181 #ifdef CONFIG_ECORE_SRIOV
182 #ifndef LINUX_REMOVE
183 /**
184 * @brief mark/clear all VFs before/after an incoming PCIe sriov
185 * disable.
186 *
187 * @param p_dev
188 * @param to_disable
189 */
190 void ecore_iov_set_vfs_to_disable(struct ecore_dev *p_dev,
191 u8 to_disable);
192
193 /**
194 * @brief mark/clear chosen VF before/after an incoming PCIe
195 * sriov disable.
196 *
197 * @param p_dev
198 * @param rel_vf_id
199 * @param to_disable
200 */
201 void ecore_iov_set_vf_to_disable(struct ecore_dev *p_dev,
202 u16 rel_vf_id,
203 u8 to_disable);
204
205 /**
206 * @brief ecore_iov_init_hw_for_vf - initialize the HW for
207 * enabling access of a VF. Also includes preparing the
208 * IGU for VF access. This needs to be called AFTER hw is
209 * initialized and BEFORE VF is loaded inside the VM.
210 *
211 * @param p_hwfn
212 * @param p_ptt
213 * @param p_params
214 *
215 * @return enum _ecore_status_t
216 */
217 enum _ecore_status_t
218 ecore_iov_init_hw_for_vf(struct ecore_hwfn *p_hwfn,
219 struct ecore_ptt *p_ptt,
220 struct ecore_iov_vf_init_params *p_params);
221
222 /**
223 * @brief ecore_iov_process_mbx_req - process a request received
224 * from the VF
225 *
226 * @param p_hwfn
227 * @param p_ptt
228 * @param vfid
229 */
230 void ecore_iov_process_mbx_req(struct ecore_hwfn *p_hwfn,
231 struct ecore_ptt *p_ptt,
232 int vfid);
233
234 /**
235 * @brief ecore_iov_release_hw_for_vf - called once upper layer
236 * knows VF is done with - can release any resources
237 * allocated for VF at this point. this must be done once
238 * we know VF is no longer loaded in VM.
239 *
240 * @param p_hwfn
241 * @param p_ptt
242 * @param rel_vf_id
243 *
244 * @return enum _ecore_status_t
245 */
246 enum _ecore_status_t ecore_iov_release_hw_for_vf(struct ecore_hwfn *p_hwfn,
247 struct ecore_ptt *p_ptt,
248 u16 rel_vf_id);
249
250 /**
251 * @brief ecore_iov_set_vf_ctx - set a context for a given VF
252 *
253 * @param p_hwfn
254 * @param vf_id
255 * @param ctx
256 *
257 * @return enum _ecore_status_t
258 */
259 enum _ecore_status_t ecore_iov_set_vf_ctx(struct ecore_hwfn *p_hwfn,
260 u16 vf_id,
261 void *ctx);
262
263 /**
264 * @brief FLR cleanup for all VFs
265 *
266 * @param p_hwfn
267 * @param p_ptt
268 *
269 * @return enum _ecore_status_t
270 */
271 enum _ecore_status_t ecore_iov_vf_flr_cleanup(struct ecore_hwfn *p_hwfn,
272 struct ecore_ptt *p_ptt);
273
274 /**
275 * @brief FLR cleanup for single VF
276 *
277 * @param p_hwfn
278 * @param p_ptt
279 * @param rel_vf_id
280 *
281 * @return enum _ecore_status_t
282 */
283 enum _ecore_status_t
284 ecore_iov_single_vf_flr_cleanup(struct ecore_hwfn *p_hwfn,
285 struct ecore_ptt *p_ptt,
286 u16 rel_vf_id);
287
288 /**
289 * @brief Update the bulletin with link information. Notice this does NOT
290 * send a bulletin update, only updates the PF's bulletin.
291 *
292 * @param p_hwfn
293 * @param p_vf
294 * @param params - the link params to use for the VF link configuration
295 * @param link - the link output to use for the VF link configuration
296 * @param p_caps - the link default capabilities.
297 */
298 void ecore_iov_set_link(struct ecore_hwfn *p_hwfn,
299 u16 vfid,
300 struct ecore_mcp_link_params *params,
301 struct ecore_mcp_link_state *link,
302 struct ecore_mcp_link_capabilities *p_caps);
303
304 /**
305 * @brief Returns link information as perceived by VF.
306 *
307 * @param p_hwfn
308 * @param p_vf
309 * @param p_params - the link params visible to vf.
310 * @param p_link - the link state visible to vf.
311 * @param p_caps - the link default capabilities visible to vf.
312 */
313 void ecore_iov_get_link(struct ecore_hwfn *p_hwfn,
314 u16 vfid,
315 struct ecore_mcp_link_params *params,
316 struct ecore_mcp_link_state *link,
317 struct ecore_mcp_link_capabilities *p_caps);
318
319 /**
320 * @brief return if the VF is pending FLR
321 *
322 * @param p_hwfn
323 * @param rel_vf_id
324 *
325 * @return bool
326 */
327 bool ecore_iov_is_vf_pending_flr(struct ecore_hwfn *p_hwfn,
328 u16 rel_vf_id);
329 #endif
330
331 /**
332 * @brief Check if given VF ID @vfid is valid
333 * w.r.t. @b_enabled_only value
334 * if b_enabled_only = true - only enabled VF id is valid
335 * else any VF id less than max_vfs is valid
336 *
337 * @param p_hwfn
338 * @param rel_vf_id - Relative VF ID
339 * @param b_enabled_only - consider only enabled VF
340 * @param b_non_malicious - true iff we want to validate vf isn't malicious.
341 *
342 * @return bool - true for valid VF ID
343 */
344 bool ecore_iov_is_valid_vfid(struct ecore_hwfn *p_hwfn,
345 int rel_vf_id,
346 bool b_enabled_only, bool b_non_malicious);
347
348 #ifndef LINUX_REMOVE
349 /**
350 * @brief Get VF's public info structure
351 *
352 * @param p_hwfn
353 * @param vfid - Relative VF ID
354 * @param b_enabled_only - false if want to access even if vf is disabled
355 *
356 * @return struct ecore_public_vf_info *
357 */
358 struct ecore_public_vf_info*
359 ecore_iov_get_public_vf_info(struct ecore_hwfn *p_hwfn,
360 u16 vfid, bool b_enabled_only);
361
362 /**
363 * @brief fills a bitmask of all VFs which have pending unhandled
364 * messages.
365 *
366 * @param p_hwfn
367 */
368 void ecore_iov_pf_get_pending_events(struct ecore_hwfn *p_hwfn,
369 u64 *events);
370
371 /**
372 * @brief Copy VF's message to PF's buffer
373 *
374 * @param p_hwfn
375 * @param ptt
376 * @param vfid
377 *
378 * @return enum _ecore_status_t
379 */
380 enum _ecore_status_t ecore_iov_copy_vf_msg(struct ecore_hwfn *p_hwfn,
381 struct ecore_ptt *ptt,
382 int vfid);
383 /**
384 * @brief Set forced MAC address in PFs copy of bulletin board
385 * and configures FW/HW to support the configuration.
386 *
387 * @param p_hwfn
388 * @param mac
389 * @param vfid
390 */
391 void ecore_iov_bulletin_set_forced_mac(struct ecore_hwfn *p_hwfn,
392 u8 *mac, int vfid);
393
394 /**
395 * @brief Set MAC address in PFs copy of bulletin board without
396 * configuring FW/HW.
397 *
398 * @param p_hwfn
399 * @param mac
400 * @param vfid
401 */
402 enum _ecore_status_t ecore_iov_bulletin_set_mac(struct ecore_hwfn *p_hwfn,
403 u8 *mac, int vfid);
404
405 /**
406 * @brief Set default behaviour of VF in case no vlans are configured for it
407 * whether to accept only untagged traffic or all.
408 * Must be called prior to the VF vport-start.
409 *
410 * @param p_hwfn
411 * @param b_untagged_only
412 * @param vfid
413 *
414 * @return ECORE_SUCCESS if configuration would stick.
415 */
416 enum _ecore_status_t
417 ecore_iov_bulletin_set_forced_untagged_default(struct ecore_hwfn *p_hwfn,
418 bool b_untagged_only,
419 int vfid);
420
421 /**
422 * @brief Get VFs opaque fid.
423 *
424 * @param p_hwfn
425 * @param vfid
426 * @param opaque_fid
427 */
428 void ecore_iov_get_vfs_opaque_fid(struct ecore_hwfn *p_hwfn, int vfid,
429 u16 *opaque_fid);
430
431 /**
432 * @brief Set forced VLAN [pvid] in PFs copy of bulletin board
433 * and configures FW/HW to support the configuration.
434 * Setting of pvid 0 would clear the feature.
435 * @param p_hwfn
436 * @param pvid
437 * @param vfid
438 */
439 void ecore_iov_bulletin_set_forced_vlan(struct ecore_hwfn *p_hwfn,
440 u16 pvid, int vfid);
441
442 /**
443 * @brief Check if VF has VPORT instance. This can be used
444 * to check if VPORT is active.
445 *
446 * @param p_hwfn
447 */
448 bool ecore_iov_vf_has_vport_instance(struct ecore_hwfn *p_hwfn, int vfid);
449
450 /**
451 * @brief PF posts the bulletin to the VF
452 *
453 * @param p_hwfn
454 * @param p_vf
455 * @param p_ptt
456 *
457 * @return enum _ecore_status_t
458 */
459 enum _ecore_status_t ecore_iov_post_vf_bulletin(struct ecore_hwfn *p_hwfn,
460 int vfid,
461 struct ecore_ptt *p_ptt);
462
463 /**
464 * @brief Check if given VF (@vfid) is marked as stopped
465 *
466 * @param p_hwfn
467 * @param vfid
468 *
469 * @return bool : true if stopped
470 */
471 bool ecore_iov_is_vf_stopped(struct ecore_hwfn *p_hwfn, int vfid);
472
473 /**
474 * @brief Configure VF anti spoofing
475 *
476 * @param p_hwfn
477 * @param vfid
478 * @param val - spoofchk value - true/false
479 *
480 * @return enum _ecore_status_t
481 */
482 enum _ecore_status_t ecore_iov_spoofchk_set(struct ecore_hwfn *p_hwfn,
483 int vfid, bool val);
484
485 /**
486 * @brief Get VF's configured spoof value.
487 *
488 * @param p_hwfn
489 * @param vfid
490 *
491 * @return bool - spoofchk value - true/false
492 */
493 bool ecore_iov_spoofchk_get(struct ecore_hwfn *p_hwfn, int vfid);
494
495 /**
496 * @brief Check for SRIOV sanity by PF.
497 *
498 * @param p_hwfn
499 * @param vfid
500 *
501 * @return bool - true if sanity checks passes, else false
502 */
503 bool ecore_iov_pf_sanity_check(struct ecore_hwfn *p_hwfn, int vfid);
504
505 /**
506 * @brief Get the num of VF chains.
507 *
508 * @param p_hwfn
509 *
510 * @return u8
511 */
512 u8 ecore_iov_vf_chains_per_pf(struct ecore_hwfn *p_hwfn);
513
514 /**
515 * @brief Get vf request mailbox params
516 *
517 * @param p_hwfn
518 * @param rel_vf_id
519 * @param pp_req_virt_addr
520 * @param p_req_virt_size
521 */
522 void ecore_iov_get_vf_req_virt_mbx_params(struct ecore_hwfn *p_hwfn,
523 u16 rel_vf_id,
524 void **pp_req_virt_addr,
525 u16 *p_req_virt_size);
526
527 /**
528 * @brief Get vf mailbox params
529 *
530 * @param p_hwfn
531 * @param rel_vf_id
532 * @param pp_reply_virt_addr
533 * @param p_reply_virt_size
534 */
535 void ecore_iov_get_vf_reply_virt_mbx_params(struct ecore_hwfn *p_hwfn,
536 u16 rel_vf_id,
537 void **pp_reply_virt_addr,
538 u16 *p_reply_virt_size);
539
540 /**
541 * @brief Validate if the given length is a valid vfpf message
542 * length
543 *
544 * @param length
545 *
546 * @return bool
547 */
548 bool ecore_iov_is_valid_vfpf_msg_length(u32 length);
549
550 /**
551 * @brief Return the max pfvf message length
552 *
553 * @return u32
554 */
555 u32 ecore_iov_pfvf_msg_length(void);
556
557 /**
558 * @brief Returns forced MAC address if one is configured
559 *
560 * @parm p_hwfn
561 * @parm rel_vf_id
562 *
563 * @return OSAL_NULL if mac isn't forced; Otherwise, returns MAC.
564 */
565 u8 *ecore_iov_bulletin_get_forced_mac(struct ecore_hwfn *p_hwfn,
566 u16 rel_vf_id);
567
568 /**
569 * @brief Returns pvid if one is configured
570 *
571 * @parm p_hwfn
572 * @parm rel_vf_id
573 *
574 * @return 0 if no pvid is configured, otherwise the pvid.
575 */
576 u16 ecore_iov_bulletin_get_forced_vlan(struct ecore_hwfn *p_hwfn,
577 u16 rel_vf_id);
578 /**
579 * @brief Configure VFs tx rate
580 *
581 * @param p_hwfn
582 * @param p_ptt
583 * @param vfid
584 * @param val - tx rate value in Mb/sec.
585 *
586 * @return enum _ecore_status_t
587 */
588 enum _ecore_status_t ecore_iov_configure_tx_rate(struct ecore_hwfn *p_hwfn,
589 struct ecore_ptt *p_ptt,
590 int vfid, int val);
591
592 /**
593 * @brief - Retrieves the statistics associated with a VF
594 *
595 * @param p_hwfn
596 * @param p_ptt
597 * @param vfid
598 * @param p_stats - this will be filled with the VF statistics
599 *
600 * @return ECORE_SUCCESS iff statistics were retrieved. Error otherwise.
601 */
602 enum _ecore_status_t ecore_iov_get_vf_stats(struct ecore_hwfn *p_hwfn,
603 struct ecore_ptt *p_ptt,
604 int vfid,
605 struct ecore_eth_stats *p_stats);
606
607 /**
608 * @brief - Retrieves num of rxqs chains
609 *
610 * @param p_hwfn
611 * @param rel_vf_id
612 *
613 * @return num of rxqs chains.
614 */
615 u8 ecore_iov_get_vf_num_rxqs(struct ecore_hwfn *p_hwfn,
616 u16 rel_vf_id);
617
618 /**
619 * @brief - Retrieves num of active rxqs chains
620 *
621 * @param p_hwfn
622 * @param rel_vf_id
623 *
624 * @return
625 */
626 u8 ecore_iov_get_vf_num_active_rxqs(struct ecore_hwfn *p_hwfn,
627 u16 rel_vf_id);
628
629 /**
630 * @brief - Retrieves ctx pointer
631 *
632 * @param p_hwfn
633 * @param rel_vf_id
634 *
635 * @return
636 */
637 void *ecore_iov_get_vf_ctx(struct ecore_hwfn *p_hwfn,
638 u16 rel_vf_id);
639
640 /**
641 * @brief - Retrieves VF`s num sbs
642 *
643 * @param p_hwfn
644 * @param rel_vf_id
645 *
646 * @return
647 */
648 u8 ecore_iov_get_vf_num_sbs(struct ecore_hwfn *p_hwfn,
649 u16 rel_vf_id);
650
651 /**
652 * @brief - Returm true if VF is waiting for acquire
653 *
654 * @param p_hwfn
655 * @param rel_vf_id
656 *
657 * @return
658 */
659 bool ecore_iov_is_vf_wait_for_acquire(struct ecore_hwfn *p_hwfn,
660 u16 rel_vf_id);
661
662 /**
663 * @brief - Returm true if VF is acquired but not initialized
664 *
665 * @param p_hwfn
666 * @param rel_vf_id
667 *
668 * @return
669 */
670 bool ecore_iov_is_vf_acquired_not_initialized(struct ecore_hwfn *p_hwfn,
671 u16 rel_vf_id);
672
673 /**
674 * @brief - Returm true if VF is acquired and initialized
675 *
676 * @param p_hwfn
677 * @param rel_vf_id
678 *
679 * @return
680 */
681 bool ecore_iov_is_vf_initialized(struct ecore_hwfn *p_hwfn,
682 u16 rel_vf_id);
683
684 /**
685 * @brief - Returm true if VF has started in FW
686 *
687 * @param p_hwfn
688 * @param rel_vf_id
689 *
690 * @return
691 */
692 bool ecore_iov_is_vf_started(struct ecore_hwfn *p_hwfn,
693 u16 rel_vf_id);
694
695 /**
696 * @brief - Get VF's vport min rate configured.
697 * @param p_hwfn
698 * @param rel_vf_id
699 *
700 * @return - rate in Mbps
701 */
702 int ecore_iov_get_vf_min_rate(struct ecore_hwfn *p_hwfn, int vfid);
703
704 /**
705 * @brief - Configure min rate for VF's vport.
706 * @param p_dev
707 * @param vfid
708 * @param - rate in Mbps
709 *
710 * @return
711 */
712 enum _ecore_status_t ecore_iov_configure_min_tx_rate(struct ecore_dev *p_dev,
713 int vfid, u32 rate);
714
715 #endif
716
717 /**
718 * @brief ecore_pf_configure_vf_queue_coalesce - PF configure coalesce parameters
719 * of VFs for Rx and Tx queue.
720 * While the API allows setting coalescing per-qid, all queues sharing a SB
721 * should be in same range [i.e., either 0-0x7f, 0x80-0xff or 0x100-0x1ff]
722 * otherwise configuration would break.
723 *
724 * @param p_hwfn
725 * @param rx_coal - Rx Coalesce value in micro seconds.
726 * @param tx_coal - TX Coalesce value in micro seconds.
727 * @param vf_id
728 * @param qid
729 *
730 * @return int
731 **/
732 enum _ecore_status_t
733 ecore_iov_pf_configure_vf_queue_coalesce(struct ecore_hwfn *p_hwfn,
734 u16 rx_coal, u16 tx_coal,
735 u16 vf_id, u16 qid);
736
737 /**
738 * @brief - Given a VF index, return index of next [including that] active VF.
739 *
740 * @param p_hwfn
741 * @param rel_vf_id
742 *
743 * @return MAX_NUM_VFS_E4 in case no further active VFs, otherwise index.
744 */
745 u16 ecore_iov_get_next_active_vf(struct ecore_hwfn *p_hwfn, u16 rel_vf_id);
746 void ecore_iov_bulletin_set_udp_ports(struct ecore_hwfn *p_hwfn, int vfid,
747 u16 vxlan_port, u16 geneve_port);
748
749 #ifdef CONFIG_ECORE_SW_CHANNEL
750 /**
751 * @brief Set whether PF should communicate with VF using SW/HW channel
752 * Needs to be called for an enabled VF before acquire is over
753 * [latest good point for doing that is OSAL_IOV_VF_ACQUIRE()]
754 *
755 * @param p_hwfn
756 * @param vfid - relative vf index
757 * @param b_is_hw - true iff PF is to use HW channel for communication
758 */
759 void ecore_iov_set_vf_hw_channel(struct ecore_hwfn *p_hwfn, int vfid,
760 bool b_is_hw);
761 #endif
762 #else
763 #ifndef LINUX_REMOVE
764 static OSAL_INLINE void ecore_iov_set_vfs_to_disable(struct ecore_dev OSAL_UNUSED *p_dev, u8 OSAL_UNUSED to_disable) {}
765 static OSAL_INLINE void ecore_iov_set_vf_to_disable(struct ecore_dev OSAL_UNUSED *p_dev, u16 OSAL_UNUSED rel_vf_id, u8 OSAL_UNUSED to_disable) {}
766 static OSAL_INLINE enum _ecore_status_t ecore_iov_init_hw_for_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, struct ecore_iov_vf_init_params OSAL_UNUSED *p_params) {return ECORE_INVAL;}
767 static OSAL_INLINE void ecore_iov_process_mbx_req(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid) {}
768 static OSAL_INLINE enum _ecore_status_t ecore_iov_release_hw_for_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, u16 OSAL_UNUSED rel_vf_id) {return ECORE_SUCCESS;}
769 static OSAL_INLINE enum _ecore_status_t ecore_iov_set_vf_ctx(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vf_id, OSAL_UNUSED void *ctx) {return ECORE_INVAL;}
770 static OSAL_INLINE enum _ecore_status_t ecore_iov_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt) {return ECORE_INVAL;}
771 static OSAL_INLINE enum _ecore_status_t ecore_iov_single_vf_flr_cleanup(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, u16 OSAL_UNUSED rel_vf_id) {return ECORE_INVAL;}
772 static OSAL_INLINE void ecore_iov_set_link(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, struct ecore_mcp_link_params OSAL_UNUSED *params, struct ecore_mcp_link_state OSAL_UNUSED *link, struct ecore_mcp_link_capabilities OSAL_UNUSED *p_caps) {}
773 static OSAL_INLINE void ecore_iov_get_link(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, struct ecore_mcp_link_params OSAL_UNUSED *params, struct ecore_mcp_link_state OSAL_UNUSED *link, struct ecore_mcp_link_capabilities OSAL_UNUSED *p_caps) {}
774 static OSAL_INLINE bool ecore_iov_is_vf_pending_flr(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
775 #endif
776 static OSAL_INLINE bool
777 ecore_iov_is_valid_vfid(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED rel_vf_id,
778 bool OSAL_UNUSED b_enabled_only,
779 bool OSAL_UNUSED b_non_malicious)
780 {
781 return false;
782 }
783 #ifndef LINUX_REMOVE
784 static OSAL_INLINE struct ecore_public_vf_info* ecore_iov_get_public_vf_info(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED vfid, bool OSAL_UNUSED b_enabled_only) {return OSAL_NULL;}
785 static OSAL_INLINE void ecore_iov_pf_add_pending_events(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED vfid) {}
786 static OSAL_INLINE void ecore_iov_pf_get_and_clear_pending_events(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u64 OSAL_UNUSED *events) {}
787 static OSAL_INLINE enum _ecore_status_t ecore_iov_copy_vf_msg(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *ptt, int OSAL_UNUSED vfid) {return ECORE_INVAL;}
788 static OSAL_INLINE void ecore_iov_bulletin_set_forced_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED *mac, int OSAL_UNUSED vfid) {}
789 static OSAL_INLINE enum _ecore_status_t ecore_iov_bulletin_set_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u8 OSAL_UNUSED *mac, OSAL_UNUSED int vfid) {return ECORE_INVAL;}
790 static OSAL_INLINE enum _ecore_status_t ecore_iov_bulletin_set_forced_untagged_default(struct ecore_hwfn OSAL_UNUSED *p_hwfn, bool OSAL_UNUSED b_untagged_only, int OSAL_UNUSED vfid) {return ECORE_INVAL;}
791 static OSAL_INLINE void ecore_iov_get_vfs_opaque_fid(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, u16 OSAL_UNUSED *opaque_fid) {}
792 static OSAL_INLINE void ecore_iov_bulletin_set_forced_vlan(struct ecore_hwfn OSAL_UNUSED p_hwfn, u16 OSAL_UNUSED pvid, int OSAL_UNUSED vfid) {}
793 static OSAL_INLINE bool ecore_iov_vf_has_vport_instance(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
794 static OSAL_INLINE enum _ecore_status_t ecore_iov_post_vf_bulletin(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, struct ecore_ptt OSAL_UNUSED *p_ptt) {return ECORE_INVAL;}
795 static OSAL_INLINE bool ecore_iov_is_vf_stopped(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
796 static OSAL_INLINE enum _ecore_status_t ecore_iov_spoofchk_set(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, bool OSAL_UNUSED val) {return ECORE_INVAL;}
797 static OSAL_INLINE bool ecore_iov_spoofchk_get(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
798 static OSAL_INLINE bool ecore_iov_pf_sanity_check(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) {return false;}
799 static OSAL_INLINE u8 ecore_iov_vf_chains_per_pf(struct ecore_hwfn OSAL_UNUSED *p_hwfn) {return 0;}
800 static OSAL_INLINE void ecore_iov_get_vf_req_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id, void OSAL_UNUSED **pp_req_virt_addr, u16 OSAL_UNUSED *p_req_virt_size) {}
801 static OSAL_INLINE void ecore_iov_get_vf_reply_virt_mbx_params(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id, void OSAL_UNUSED **pp_reply_virt_addr, u16 OSAL_UNUSED *p_reply_virt_size) {}
802 static OSAL_INLINE bool ecore_iov_is_valid_vfpf_msg_length(u32 OSAL_UNUSED length) {return false;}
803 static OSAL_INLINE u32 ecore_iov_pfvf_msg_length(void) {return 0;}
804 static OSAL_INLINE u8 *ecore_iov_bulletin_get_forced_mac(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return OSAL_NULL;}
805 static OSAL_INLINE u16 ecore_iov_bulletin_get_forced_vlan(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
806 static OSAL_INLINE enum _ecore_status_t ecore_iov_configure_tx_rate(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid, int OSAL_UNUSED val) { return ECORE_INVAL; }
807 static OSAL_INLINE enum _ecore_status_t ecore_iov_get_vf_stats(struct ecore_hwfn OSAL_UNUSED *p_hwfn, struct ecore_ptt OSAL_UNUSED *p_ptt, int OSAL_UNUSED vfid, struct ecore_eth_stats OSAL_UNUSED *p_stats) { return ECORE_INVAL; }
808 static OSAL_INLINE u8 ecore_iov_get_vf_num_rxqs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
809 static OSAL_INLINE u8 ecore_iov_get_vf_num_active_rxqs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
810 static OSAL_INLINE void *ecore_iov_get_vf_ctx(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return OSAL_NULL;}
811 static OSAL_INLINE u8 ecore_iov_get_vf_num_sbs(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return 0;}
812 static OSAL_INLINE bool ecore_iov_is_vf_wait_for_acquire(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
813 static OSAL_INLINE bool ecore_iov_is_vf_acquired_not_initialized(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
814 static OSAL_INLINE bool ecore_iov_is_vf_initialized(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) {return false;}
815 static OSAL_INLINE int ecore_iov_get_vf_min_rate(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid) { return 0; }
816 static OSAL_INLINE enum _ecore_status_t ecore_iov_configure_min_tx_rate(struct ecore_dev OSAL_UNUSED *p_dev, int OSAL_UNUSED vfid, OSAL_UNUSED u32 rate) { return ECORE_INVAL; }
817 #endif
818 static OSAL_INLINE void ecore_iov_bulletin_set_udp_ports(struct ecore_hwfn OSAL_UNUSED *p_hwfn, int OSAL_UNUSED vfid, u16 OSAL_UNUSED vxlan_port, u16 OSAL_UNUSED geneve_port) { return; }
819 static OSAL_INLINE u16 ecore_iov_get_next_active_vf(struct ecore_hwfn OSAL_UNUSED *p_hwfn, u16 OSAL_UNUSED rel_vf_id) { return MAX_NUM_VFS_E4; }
820
821 #ifdef CONFIG_ECORE_SW_CHANNEL
822 static OSAL_INLINE void
823 ecore_iov_set_vf_hw_channel(struct ecore_hwfn OSAL_UNUSED *p_hwfn,
824 int OSAL_UNUSED vfid, bool OSAL_UNUSED b_is_hw) {}
825 #endif
826 #endif
827
828 #define ecore_for_each_vf(_p_hwfn, _i) \
829 for (_i = ecore_iov_get_next_active_vf(_p_hwfn, 0); \
830 _i < MAX_NUM_VFS_E4; \
831 _i = ecore_iov_get_next_active_vf(_p_hwfn, _i + 1))
832
833 #endif
Cache object: acd3f11806a0736bb8d6cc0895cdd92e
|