qdisc.c (22289B)
1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause) 2/* Copyright (C) 2018 Netronome Systems, Inc. */ 3 4#include <linux/rtnetlink.h> 5#include <net/pkt_cls.h> 6#include <net/pkt_sched.h> 7#include <net/red.h> 8 9#include "../nfpcore/nfp_cpp.h" 10#include "../nfp_app.h" 11#include "../nfp_main.h" 12#include "../nfp_net.h" 13#include "../nfp_port.h" 14#include "main.h" 15 16static bool nfp_abm_qdisc_is_red(struct nfp_qdisc *qdisc) 17{ 18 return qdisc->type == NFP_QDISC_RED || qdisc->type == NFP_QDISC_GRED; 19} 20 21static bool nfp_abm_qdisc_child_valid(struct nfp_qdisc *qdisc, unsigned int id) 22{ 23 return qdisc->children[id] && 24 qdisc->children[id] != NFP_QDISC_UNTRACKED; 25} 26 27static void *nfp_abm_qdisc_tree_deref_slot(void __rcu **slot) 28{ 29 return rtnl_dereference(*slot); 30} 31 32static void 33nfp_abm_stats_propagate(struct nfp_alink_stats *parent, 34 struct nfp_alink_stats *child) 35{ 36 parent->tx_pkts += child->tx_pkts; 37 parent->tx_bytes += child->tx_bytes; 38 parent->backlog_pkts += child->backlog_pkts; 39 parent->backlog_bytes += child->backlog_bytes; 40 parent->overlimits += child->overlimits; 41 parent->drops += child->drops; 42} 43 44static void 45nfp_abm_stats_update_red(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc, 46 unsigned int queue) 47{ 48 struct nfp_cpp *cpp = alink->abm->app->cpp; 49 unsigned int i; 50 int err; 51 52 if (!qdisc->offloaded) 53 return; 54 55 for (i = 0; i < qdisc->red.num_bands; i++) { 56 err = nfp_abm_ctrl_read_q_stats(alink, i, queue, 57 &qdisc->red.band[i].stats); 58 if (err) 59 nfp_err(cpp, "RED stats (%d, %d) read failed with error %d\n", 60 i, queue, err); 61 62 err = nfp_abm_ctrl_read_q_xstats(alink, i, queue, 63 &qdisc->red.band[i].xstats); 64 if (err) 65 nfp_err(cpp, "RED xstats (%d, %d) read failed with error %d\n", 66 i, queue, err); 67 } 68} 69 70static void 71nfp_abm_stats_update_mq(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) 72{ 73 unsigned int i; 74 75 if (qdisc->type != NFP_QDISC_MQ) 76 return; 77 78 for (i = 0; i < alink->total_queues; i++) 79 if (nfp_abm_qdisc_child_valid(qdisc, i)) 80 nfp_abm_stats_update_red(alink, qdisc->children[i], i); 81} 82 83static void __nfp_abm_stats_update(struct nfp_abm_link *alink, u64 time_now) 84{ 85 alink->last_stats_update = time_now; 86 if (alink->root_qdisc) 87 nfp_abm_stats_update_mq(alink, alink->root_qdisc); 88} 89 90static void nfp_abm_stats_update(struct nfp_abm_link *alink) 91{ 92 u64 now; 93 94 /* Limit the frequency of updates - stats of non-leaf qdiscs are a sum 95 * of all their leafs, so we would read the same stat multiple times 96 * for every dump. 97 */ 98 now = ktime_get(); 99 if (now - alink->last_stats_update < NFP_ABM_STATS_REFRESH_IVAL) 100 return; 101 102 __nfp_abm_stats_update(alink, now); 103} 104 105static void 106nfp_abm_qdisc_unlink_children(struct nfp_qdisc *qdisc, 107 unsigned int start, unsigned int end) 108{ 109 unsigned int i; 110 111 for (i = start; i < end; i++) 112 if (nfp_abm_qdisc_child_valid(qdisc, i)) { 113 qdisc->children[i]->use_cnt--; 114 qdisc->children[i] = NULL; 115 } 116} 117 118static void 119nfp_abm_qdisc_offload_stop(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) 120{ 121 unsigned int i; 122 123 /* Don't complain when qdisc is getting unlinked */ 124 if (qdisc->use_cnt) 125 nfp_warn(alink->abm->app->cpp, "Offload of '%08x' stopped\n", 126 qdisc->handle); 127 128 if (!nfp_abm_qdisc_is_red(qdisc)) 129 return; 130 131 for (i = 0; i < qdisc->red.num_bands; i++) { 132 qdisc->red.band[i].stats.backlog_pkts = 0; 133 qdisc->red.band[i].stats.backlog_bytes = 0; 134 } 135} 136 137static int 138__nfp_abm_stats_init(struct nfp_abm_link *alink, unsigned int band, 139 unsigned int queue, struct nfp_alink_stats *prev_stats, 140 struct nfp_alink_xstats *prev_xstats) 141{ 142 u64 backlog_pkts, backlog_bytes; 143 int err; 144 145 /* Don't touch the backlog, backlog can only be reset after it has 146 * been reported back to the tc qdisc stats. 147 */ 148 backlog_pkts = prev_stats->backlog_pkts; 149 backlog_bytes = prev_stats->backlog_bytes; 150 151 err = nfp_abm_ctrl_read_q_stats(alink, band, queue, prev_stats); 152 if (err) { 153 nfp_err(alink->abm->app->cpp, 154 "RED stats init (%d, %d) failed with error %d\n", 155 band, queue, err); 156 return err; 157 } 158 159 err = nfp_abm_ctrl_read_q_xstats(alink, band, queue, prev_xstats); 160 if (err) { 161 nfp_err(alink->abm->app->cpp, 162 "RED xstats init (%d, %d) failed with error %d\n", 163 band, queue, err); 164 return err; 165 } 166 167 prev_stats->backlog_pkts = backlog_pkts; 168 prev_stats->backlog_bytes = backlog_bytes; 169 return 0; 170} 171 172static int 173nfp_abm_stats_init(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc, 174 unsigned int queue) 175{ 176 unsigned int i; 177 int err; 178 179 for (i = 0; i < qdisc->red.num_bands; i++) { 180 err = __nfp_abm_stats_init(alink, i, queue, 181 &qdisc->red.band[i].prev_stats, 182 &qdisc->red.band[i].prev_xstats); 183 if (err) 184 return err; 185 } 186 187 return 0; 188} 189 190static void 191nfp_abm_offload_compile_red(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc, 192 unsigned int queue) 193{ 194 bool good_red, good_gred; 195 unsigned int i; 196 197 good_red = qdisc->type == NFP_QDISC_RED && 198 qdisc->params_ok && 199 qdisc->use_cnt == 1 && 200 !alink->has_prio && 201 !qdisc->children[0]; 202 good_gred = qdisc->type == NFP_QDISC_GRED && 203 qdisc->params_ok && 204 qdisc->use_cnt == 1; 205 qdisc->offload_mark = good_red || good_gred; 206 207 /* If we are starting offload init prev_stats */ 208 if (qdisc->offload_mark && !qdisc->offloaded) 209 if (nfp_abm_stats_init(alink, qdisc, queue)) 210 qdisc->offload_mark = false; 211 212 if (!qdisc->offload_mark) 213 return; 214 215 for (i = 0; i < alink->abm->num_bands; i++) { 216 enum nfp_abm_q_action act; 217 218 nfp_abm_ctrl_set_q_lvl(alink, i, queue, 219 qdisc->red.band[i].threshold); 220 act = qdisc->red.band[i].ecn ? 221 NFP_ABM_ACT_MARK_DROP : NFP_ABM_ACT_DROP; 222 nfp_abm_ctrl_set_q_act(alink, i, queue, act); 223 } 224} 225 226static void 227nfp_abm_offload_compile_mq(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) 228{ 229 unsigned int i; 230 231 qdisc->offload_mark = qdisc->type == NFP_QDISC_MQ; 232 if (!qdisc->offload_mark) 233 return; 234 235 for (i = 0; i < alink->total_queues; i++) { 236 struct nfp_qdisc *child = qdisc->children[i]; 237 238 if (!nfp_abm_qdisc_child_valid(qdisc, i)) 239 continue; 240 241 nfp_abm_offload_compile_red(alink, child, i); 242 } 243} 244 245void nfp_abm_qdisc_offload_update(struct nfp_abm_link *alink) 246{ 247 struct nfp_abm *abm = alink->abm; 248 struct radix_tree_iter iter; 249 struct nfp_qdisc *qdisc; 250 void __rcu **slot; 251 size_t i; 252 253 /* Mark all thresholds as unconfigured */ 254 for (i = 0; i < abm->num_bands; i++) 255 __bitmap_set(abm->threshold_undef, 256 i * NFP_NET_MAX_RX_RINGS + alink->queue_base, 257 alink->total_queues); 258 259 /* Clear offload marks */ 260 radix_tree_for_each_slot(slot, &alink->qdiscs, &iter, 0) { 261 qdisc = nfp_abm_qdisc_tree_deref_slot(slot); 262 qdisc->offload_mark = false; 263 } 264 265 if (alink->root_qdisc) 266 nfp_abm_offload_compile_mq(alink, alink->root_qdisc); 267 268 /* Refresh offload status */ 269 radix_tree_for_each_slot(slot, &alink->qdiscs, &iter, 0) { 270 qdisc = nfp_abm_qdisc_tree_deref_slot(slot); 271 if (!qdisc->offload_mark && qdisc->offloaded) 272 nfp_abm_qdisc_offload_stop(alink, qdisc); 273 qdisc->offloaded = qdisc->offload_mark; 274 } 275 276 /* Reset the unconfigured thresholds */ 277 for (i = 0; i < abm->num_thresholds; i++) 278 if (test_bit(i, abm->threshold_undef)) 279 __nfp_abm_ctrl_set_q_lvl(abm, i, NFP_ABM_LVL_INFINITY); 280 281 __nfp_abm_stats_update(alink, ktime_get()); 282} 283 284static void 285nfp_abm_qdisc_clear_mq(struct net_device *netdev, struct nfp_abm_link *alink, 286 struct nfp_qdisc *qdisc) 287{ 288 struct radix_tree_iter iter; 289 unsigned int mq_refs = 0; 290 void __rcu **slot; 291 292 if (!qdisc->use_cnt) 293 return; 294 /* MQ doesn't notify well on destruction, we need special handling of 295 * MQ's children. 296 */ 297 if (qdisc->type == NFP_QDISC_MQ && 298 qdisc == alink->root_qdisc && 299 netdev->reg_state == NETREG_UNREGISTERING) 300 return; 301 302 /* Count refs held by MQ instances and clear pointers */ 303 radix_tree_for_each_slot(slot, &alink->qdiscs, &iter, 0) { 304 struct nfp_qdisc *mq = nfp_abm_qdisc_tree_deref_slot(slot); 305 unsigned int i; 306 307 if (mq->type != NFP_QDISC_MQ || mq->netdev != netdev) 308 continue; 309 for (i = 0; i < mq->num_children; i++) 310 if (mq->children[i] == qdisc) { 311 mq->children[i] = NULL; 312 mq_refs++; 313 } 314 } 315 316 WARN(qdisc->use_cnt != mq_refs, "non-zero qdisc use count: %d (- %d)\n", 317 qdisc->use_cnt, mq_refs); 318} 319 320static void 321nfp_abm_qdisc_free(struct net_device *netdev, struct nfp_abm_link *alink, 322 struct nfp_qdisc *qdisc) 323{ 324 struct nfp_port *port = nfp_port_from_netdev(netdev); 325 326 if (!qdisc) 327 return; 328 nfp_abm_qdisc_clear_mq(netdev, alink, qdisc); 329 WARN_ON(radix_tree_delete(&alink->qdiscs, 330 TC_H_MAJ(qdisc->handle)) != qdisc); 331 332 kfree(qdisc->children); 333 kfree(qdisc); 334 335 port->tc_offload_cnt--; 336} 337 338static struct nfp_qdisc * 339nfp_abm_qdisc_alloc(struct net_device *netdev, struct nfp_abm_link *alink, 340 enum nfp_qdisc_type type, u32 parent_handle, u32 handle, 341 unsigned int children) 342{ 343 struct nfp_port *port = nfp_port_from_netdev(netdev); 344 struct nfp_qdisc *qdisc; 345 int err; 346 347 qdisc = kzalloc(sizeof(*qdisc), GFP_KERNEL); 348 if (!qdisc) 349 return NULL; 350 351 if (children) { 352 qdisc->children = kcalloc(children, sizeof(void *), GFP_KERNEL); 353 if (!qdisc->children) 354 goto err_free_qdisc; 355 } 356 357 qdisc->netdev = netdev; 358 qdisc->type = type; 359 qdisc->parent_handle = parent_handle; 360 qdisc->handle = handle; 361 qdisc->num_children = children; 362 363 err = radix_tree_insert(&alink->qdiscs, TC_H_MAJ(qdisc->handle), qdisc); 364 if (err) { 365 nfp_err(alink->abm->app->cpp, 366 "Qdisc insertion into radix tree failed: %d\n", err); 367 goto err_free_child_tbl; 368 } 369 370 port->tc_offload_cnt++; 371 return qdisc; 372 373err_free_child_tbl: 374 kfree(qdisc->children); 375err_free_qdisc: 376 kfree(qdisc); 377 return NULL; 378} 379 380static struct nfp_qdisc * 381nfp_abm_qdisc_find(struct nfp_abm_link *alink, u32 handle) 382{ 383 return radix_tree_lookup(&alink->qdiscs, TC_H_MAJ(handle)); 384} 385 386static int 387nfp_abm_qdisc_replace(struct net_device *netdev, struct nfp_abm_link *alink, 388 enum nfp_qdisc_type type, u32 parent_handle, u32 handle, 389 unsigned int children, struct nfp_qdisc **qdisc) 390{ 391 *qdisc = nfp_abm_qdisc_find(alink, handle); 392 if (*qdisc) { 393 if (WARN_ON((*qdisc)->type != type)) 394 return -EINVAL; 395 return 1; 396 } 397 398 *qdisc = nfp_abm_qdisc_alloc(netdev, alink, type, parent_handle, handle, 399 children); 400 return *qdisc ? 0 : -ENOMEM; 401} 402 403static void 404nfp_abm_qdisc_destroy(struct net_device *netdev, struct nfp_abm_link *alink, 405 u32 handle) 406{ 407 struct nfp_qdisc *qdisc; 408 409 qdisc = nfp_abm_qdisc_find(alink, handle); 410 if (!qdisc) 411 return; 412 413 /* We don't get TC_SETUP_ROOT_QDISC w/ MQ when netdev is unregistered */ 414 if (alink->root_qdisc == qdisc) 415 qdisc->use_cnt--; 416 417 nfp_abm_qdisc_unlink_children(qdisc, 0, qdisc->num_children); 418 nfp_abm_qdisc_free(netdev, alink, qdisc); 419 420 if (alink->root_qdisc == qdisc) { 421 alink->root_qdisc = NULL; 422 /* Only root change matters, other changes are acted upon on 423 * the graft notification. 424 */ 425 nfp_abm_qdisc_offload_update(alink); 426 } 427} 428 429static int 430nfp_abm_qdisc_graft(struct nfp_abm_link *alink, u32 handle, u32 child_handle, 431 unsigned int id) 432{ 433 struct nfp_qdisc *parent, *child; 434 435 parent = nfp_abm_qdisc_find(alink, handle); 436 if (!parent) 437 return 0; 438 439 if (WARN(id >= parent->num_children, 440 "graft child out of bound %d >= %d\n", 441 id, parent->num_children)) 442 return -EINVAL; 443 444 nfp_abm_qdisc_unlink_children(parent, id, id + 1); 445 446 child = nfp_abm_qdisc_find(alink, child_handle); 447 if (child) 448 child->use_cnt++; 449 else 450 child = NFP_QDISC_UNTRACKED; 451 parent->children[id] = child; 452 453 nfp_abm_qdisc_offload_update(alink); 454 455 return 0; 456} 457 458static void 459nfp_abm_stats_calculate(struct nfp_alink_stats *new, 460 struct nfp_alink_stats *old, 461 struct gnet_stats_basic_sync *bstats, 462 struct gnet_stats_queue *qstats) 463{ 464 _bstats_update(bstats, new->tx_bytes - old->tx_bytes, 465 new->tx_pkts - old->tx_pkts); 466 qstats->qlen += new->backlog_pkts - old->backlog_pkts; 467 qstats->backlog += new->backlog_bytes - old->backlog_bytes; 468 qstats->overlimits += new->overlimits - old->overlimits; 469 qstats->drops += new->drops - old->drops; 470} 471 472static void 473nfp_abm_stats_red_calculate(struct nfp_alink_xstats *new, 474 struct nfp_alink_xstats *old, 475 struct red_stats *stats) 476{ 477 stats->forced_mark += new->ecn_marked - old->ecn_marked; 478 stats->pdrop += new->pdrop - old->pdrop; 479} 480 481static int 482nfp_abm_gred_stats(struct nfp_abm_link *alink, u32 handle, 483 struct tc_gred_qopt_offload_stats *stats) 484{ 485 struct nfp_qdisc *qdisc; 486 unsigned int i; 487 488 nfp_abm_stats_update(alink); 489 490 qdisc = nfp_abm_qdisc_find(alink, handle); 491 if (!qdisc) 492 return -EOPNOTSUPP; 493 /* If the qdisc offload has stopped we may need to adjust the backlog 494 * counters back so carry on even if qdisc is not currently offloaded. 495 */ 496 497 for (i = 0; i < qdisc->red.num_bands; i++) { 498 if (!stats->xstats[i]) 499 continue; 500 501 nfp_abm_stats_calculate(&qdisc->red.band[i].stats, 502 &qdisc->red.band[i].prev_stats, 503 &stats->bstats[i], &stats->qstats[i]); 504 qdisc->red.band[i].prev_stats = qdisc->red.band[i].stats; 505 506 nfp_abm_stats_red_calculate(&qdisc->red.band[i].xstats, 507 &qdisc->red.band[i].prev_xstats, 508 stats->xstats[i]); 509 qdisc->red.band[i].prev_xstats = qdisc->red.band[i].xstats; 510 } 511 512 return qdisc->offloaded ? 0 : -EOPNOTSUPP; 513} 514 515static bool 516nfp_abm_gred_check_params(struct nfp_abm_link *alink, 517 struct tc_gred_qopt_offload *opt) 518{ 519 struct nfp_cpp *cpp = alink->abm->app->cpp; 520 struct nfp_abm *abm = alink->abm; 521 unsigned int i; 522 523 if (opt->set.grio_on || opt->set.wred_on) { 524 nfp_warn(cpp, "GRED offload failed - GRIO and WRED not supported (p:%08x h:%08x)\n", 525 opt->parent, opt->handle); 526 return false; 527 } 528 if (opt->set.dp_def != alink->def_band) { 529 nfp_warn(cpp, "GRED offload failed - default band must be %d (p:%08x h:%08x)\n", 530 alink->def_band, opt->parent, opt->handle); 531 return false; 532 } 533 if (opt->set.dp_cnt != abm->num_bands) { 534 nfp_warn(cpp, "GRED offload failed - band count must be %d (p:%08x h:%08x)\n", 535 abm->num_bands, opt->parent, opt->handle); 536 return false; 537 } 538 539 for (i = 0; i < abm->num_bands; i++) { 540 struct tc_gred_vq_qopt_offload_params *band = &opt->set.tab[i]; 541 542 if (!band->present) 543 return false; 544 if (!band->is_ecn && !nfp_abm_has_drop(abm)) { 545 nfp_warn(cpp, "GRED offload failed - drop is not supported (ECN option required) (p:%08x h:%08x vq:%d)\n", 546 opt->parent, opt->handle, i); 547 return false; 548 } 549 if (band->is_ecn && !nfp_abm_has_mark(abm)) { 550 nfp_warn(cpp, "GRED offload failed - ECN marking not supported (p:%08x h:%08x vq:%d)\n", 551 opt->parent, opt->handle, i); 552 return false; 553 } 554 if (band->is_harddrop) { 555 nfp_warn(cpp, "GRED offload failed - harddrop is not supported (p:%08x h:%08x vq:%d)\n", 556 opt->parent, opt->handle, i); 557 return false; 558 } 559 if (band->min != band->max) { 560 nfp_warn(cpp, "GRED offload failed - threshold mismatch (p:%08x h:%08x vq:%d)\n", 561 opt->parent, opt->handle, i); 562 return false; 563 } 564 if (band->min > S32_MAX) { 565 nfp_warn(cpp, "GRED offload failed - threshold too large %d > %d (p:%08x h:%08x vq:%d)\n", 566 band->min, S32_MAX, opt->parent, opt->handle, 567 i); 568 return false; 569 } 570 } 571 572 return true; 573} 574 575static int 576nfp_abm_gred_replace(struct net_device *netdev, struct nfp_abm_link *alink, 577 struct tc_gred_qopt_offload *opt) 578{ 579 struct nfp_qdisc *qdisc; 580 unsigned int i; 581 int ret; 582 583 ret = nfp_abm_qdisc_replace(netdev, alink, NFP_QDISC_GRED, opt->parent, 584 opt->handle, 0, &qdisc); 585 if (ret < 0) 586 return ret; 587 588 qdisc->params_ok = nfp_abm_gred_check_params(alink, opt); 589 if (qdisc->params_ok) { 590 qdisc->red.num_bands = opt->set.dp_cnt; 591 for (i = 0; i < qdisc->red.num_bands; i++) { 592 qdisc->red.band[i].ecn = opt->set.tab[i].is_ecn; 593 qdisc->red.band[i].threshold = opt->set.tab[i].min; 594 } 595 } 596 597 if (qdisc->use_cnt) 598 nfp_abm_qdisc_offload_update(alink); 599 600 return 0; 601} 602 603int nfp_abm_setup_tc_gred(struct net_device *netdev, struct nfp_abm_link *alink, 604 struct tc_gred_qopt_offload *opt) 605{ 606 switch (opt->command) { 607 case TC_GRED_REPLACE: 608 return nfp_abm_gred_replace(netdev, alink, opt); 609 case TC_GRED_DESTROY: 610 nfp_abm_qdisc_destroy(netdev, alink, opt->handle); 611 return 0; 612 case TC_GRED_STATS: 613 return nfp_abm_gred_stats(alink, opt->handle, &opt->stats); 614 default: 615 return -EOPNOTSUPP; 616 } 617} 618 619static int 620nfp_abm_red_xstats(struct nfp_abm_link *alink, struct tc_red_qopt_offload *opt) 621{ 622 struct nfp_qdisc *qdisc; 623 624 nfp_abm_stats_update(alink); 625 626 qdisc = nfp_abm_qdisc_find(alink, opt->handle); 627 if (!qdisc || !qdisc->offloaded) 628 return -EOPNOTSUPP; 629 630 nfp_abm_stats_red_calculate(&qdisc->red.band[0].xstats, 631 &qdisc->red.band[0].prev_xstats, 632 opt->xstats); 633 qdisc->red.band[0].prev_xstats = qdisc->red.band[0].xstats; 634 return 0; 635} 636 637static int 638nfp_abm_red_stats(struct nfp_abm_link *alink, u32 handle, 639 struct tc_qopt_offload_stats *stats) 640{ 641 struct nfp_qdisc *qdisc; 642 643 nfp_abm_stats_update(alink); 644 645 qdisc = nfp_abm_qdisc_find(alink, handle); 646 if (!qdisc) 647 return -EOPNOTSUPP; 648 /* If the qdisc offload has stopped we may need to adjust the backlog 649 * counters back so carry on even if qdisc is not currently offloaded. 650 */ 651 652 nfp_abm_stats_calculate(&qdisc->red.band[0].stats, 653 &qdisc->red.band[0].prev_stats, 654 stats->bstats, stats->qstats); 655 qdisc->red.band[0].prev_stats = qdisc->red.band[0].stats; 656 657 return qdisc->offloaded ? 0 : -EOPNOTSUPP; 658} 659 660static bool 661nfp_abm_red_check_params(struct nfp_abm_link *alink, 662 struct tc_red_qopt_offload *opt) 663{ 664 struct nfp_cpp *cpp = alink->abm->app->cpp; 665 struct nfp_abm *abm = alink->abm; 666 667 if (!opt->set.is_ecn && !nfp_abm_has_drop(abm)) { 668 nfp_warn(cpp, "RED offload failed - drop is not supported (ECN option required) (p:%08x h:%08x)\n", 669 opt->parent, opt->handle); 670 return false; 671 } 672 if (opt->set.is_ecn && !nfp_abm_has_mark(abm)) { 673 nfp_warn(cpp, "RED offload failed - ECN marking not supported (p:%08x h:%08x)\n", 674 opt->parent, opt->handle); 675 return false; 676 } 677 if (opt->set.is_harddrop) { 678 nfp_warn(cpp, "RED offload failed - harddrop is not supported (p:%08x h:%08x)\n", 679 opt->parent, opt->handle); 680 return false; 681 } 682 if (opt->set.min != opt->set.max) { 683 nfp_warn(cpp, "RED offload failed - unsupported min/max parameters (p:%08x h:%08x)\n", 684 opt->parent, opt->handle); 685 return false; 686 } 687 if (opt->set.min > NFP_ABM_LVL_INFINITY) { 688 nfp_warn(cpp, "RED offload failed - threshold too large %d > %d (p:%08x h:%08x)\n", 689 opt->set.min, NFP_ABM_LVL_INFINITY, opt->parent, 690 opt->handle); 691 return false; 692 } 693 694 return true; 695} 696 697static int 698nfp_abm_red_replace(struct net_device *netdev, struct nfp_abm_link *alink, 699 struct tc_red_qopt_offload *opt) 700{ 701 struct nfp_qdisc *qdisc; 702 int ret; 703 704 ret = nfp_abm_qdisc_replace(netdev, alink, NFP_QDISC_RED, opt->parent, 705 opt->handle, 1, &qdisc); 706 if (ret < 0) 707 return ret; 708 709 /* If limit != 0 child gets reset */ 710 if (opt->set.limit) { 711 if (nfp_abm_qdisc_child_valid(qdisc, 0)) 712 qdisc->children[0]->use_cnt--; 713 qdisc->children[0] = NULL; 714 } else { 715 /* Qdisc was just allocated without a limit will use noop_qdisc, 716 * i.e. a block hole. 717 */ 718 if (!ret) 719 qdisc->children[0] = NFP_QDISC_UNTRACKED; 720 } 721 722 qdisc->params_ok = nfp_abm_red_check_params(alink, opt); 723 if (qdisc->params_ok) { 724 qdisc->red.num_bands = 1; 725 qdisc->red.band[0].ecn = opt->set.is_ecn; 726 qdisc->red.band[0].threshold = opt->set.min; 727 } 728 729 if (qdisc->use_cnt == 1) 730 nfp_abm_qdisc_offload_update(alink); 731 732 return 0; 733} 734 735int nfp_abm_setup_tc_red(struct net_device *netdev, struct nfp_abm_link *alink, 736 struct tc_red_qopt_offload *opt) 737{ 738 switch (opt->command) { 739 case TC_RED_REPLACE: 740 return nfp_abm_red_replace(netdev, alink, opt); 741 case TC_RED_DESTROY: 742 nfp_abm_qdisc_destroy(netdev, alink, opt->handle); 743 return 0; 744 case TC_RED_STATS: 745 return nfp_abm_red_stats(alink, opt->handle, &opt->stats); 746 case TC_RED_XSTATS: 747 return nfp_abm_red_xstats(alink, opt); 748 case TC_RED_GRAFT: 749 return nfp_abm_qdisc_graft(alink, opt->handle, 750 opt->child_handle, 0); 751 default: 752 return -EOPNOTSUPP; 753 } 754} 755 756static int 757nfp_abm_mq_create(struct net_device *netdev, struct nfp_abm_link *alink, 758 struct tc_mq_qopt_offload *opt) 759{ 760 struct nfp_qdisc *qdisc; 761 int ret; 762 763 ret = nfp_abm_qdisc_replace(netdev, alink, NFP_QDISC_MQ, 764 TC_H_ROOT, opt->handle, alink->total_queues, 765 &qdisc); 766 if (ret < 0) 767 return ret; 768 769 qdisc->params_ok = true; 770 qdisc->offloaded = true; 771 nfp_abm_qdisc_offload_update(alink); 772 return 0; 773} 774 775static int 776nfp_abm_mq_stats(struct nfp_abm_link *alink, u32 handle, 777 struct tc_qopt_offload_stats *stats) 778{ 779 struct nfp_qdisc *qdisc, *red; 780 unsigned int i, j; 781 782 qdisc = nfp_abm_qdisc_find(alink, handle); 783 if (!qdisc) 784 return -EOPNOTSUPP; 785 786 nfp_abm_stats_update(alink); 787 788 /* MQ stats are summed over the children in the core, so we need 789 * to add up the unreported child values. 790 */ 791 memset(&qdisc->mq.stats, 0, sizeof(qdisc->mq.stats)); 792 memset(&qdisc->mq.prev_stats, 0, sizeof(qdisc->mq.prev_stats)); 793 794 for (i = 0; i < qdisc->num_children; i++) { 795 if (!nfp_abm_qdisc_child_valid(qdisc, i)) 796 continue; 797 798 if (!nfp_abm_qdisc_is_red(qdisc->children[i])) 799 continue; 800 red = qdisc->children[i]; 801 802 for (j = 0; j < red->red.num_bands; j++) { 803 nfp_abm_stats_propagate(&qdisc->mq.stats, 804 &red->red.band[j].stats); 805 nfp_abm_stats_propagate(&qdisc->mq.prev_stats, 806 &red->red.band[j].prev_stats); 807 } 808 } 809 810 nfp_abm_stats_calculate(&qdisc->mq.stats, &qdisc->mq.prev_stats, 811 stats->bstats, stats->qstats); 812 813 return qdisc->offloaded ? 0 : -EOPNOTSUPP; 814} 815 816int nfp_abm_setup_tc_mq(struct net_device *netdev, struct nfp_abm_link *alink, 817 struct tc_mq_qopt_offload *opt) 818{ 819 switch (opt->command) { 820 case TC_MQ_CREATE: 821 return nfp_abm_mq_create(netdev, alink, opt); 822 case TC_MQ_DESTROY: 823 nfp_abm_qdisc_destroy(netdev, alink, opt->handle); 824 return 0; 825 case TC_MQ_STATS: 826 return nfp_abm_mq_stats(alink, opt->handle, &opt->stats); 827 case TC_MQ_GRAFT: 828 return nfp_abm_qdisc_graft(alink, opt->handle, 829 opt->graft_params.child_handle, 830 opt->graft_params.queue); 831 default: 832 return -EOPNOTSUPP; 833 } 834} 835 836int nfp_abm_setup_root(struct net_device *netdev, struct nfp_abm_link *alink, 837 struct tc_root_qopt_offload *opt) 838{ 839 if (opt->ingress) 840 return -EOPNOTSUPP; 841 if (alink->root_qdisc) 842 alink->root_qdisc->use_cnt--; 843 alink->root_qdisc = nfp_abm_qdisc_find(alink, opt->handle); 844 if (alink->root_qdisc) 845 alink->root_qdisc->use_cnt++; 846 847 nfp_abm_qdisc_offload_update(alink); 848 849 return 0; 850}