Searched refs:cnp (Results 1 - 12 of 12) sorted by relevance

H A Dtopo_snap.c560 step_child(tnode_t *cnp, topo_walk_t *wp, int flag, int bottomup) argument
565 nnp = topo_child_first(cnp);
570 cnp->tn_name, cnp->tn_instance);
576 cnp->tn_name, cnp->tn_instance, nnp->tn_name, nnp->tn_instance);
589 step_sibling(tnode_t *cnp, topo_walk_t *wp, int flag, int bottomup) argument
594 nnp = topo_child_next(cnp->tn_parent, cnp);
599 cnp
621 tnode_t *nnp, *cnp; local
644 tnode_t *cnp, *pnp; local
663 tnode_t *cnp = wp->tw_node; local
731 tnode_t *cnp; local
H A Dtopo_mod.c947 tnode_t *cnp; local
957 for (cnp = topo_child_first(node); cnp != NULL;
958 cnp = topo_child_next(node, cnp)) {
959 if (topo_node_flags(cnp) != TOPO_NODE_FACILITY)
H A Ddevinfo_devlink.c543 cache_node_t *cnp; local
566 cnp = node_insert(hdp, pcnp, path, INSERT_HEAD);
567 if (cnp == NULL) {
575 if (read_minors(hdp, cnp, dnp->minor) != 0 ||
576 read_nodes(hdp, cnp, dnp->child) != 0) {
581 cnp->path);
799 cache_node_t *cnp,
808 for (; cnp != NULL; cnp = cnp
796 write_nodes( struct di_devlink_handle *hdp, struct db_node *pdnp, cache_node_t *cnp, uint32_t *next) argument
1314 cache_node_t *cnp = *pp; local
1465 cache_node_t *cnp; local
1607 cache_node_t *cnp = tnp->node; local
1687 delete_unused_nodes(di_devlink_handle_t hdp, cache_node_t *cnp) argument
2650 cache_node_t *cnp; local
2982 count_node(cache_node_t *cnp, uint32_t *count) argument
3705 cache_node_t *cnp; local
H A Ddevinfo_devlink.h315 cache_node_t *cnp, uint32_t *next);
331 static void count_node(cache_node_t *cnp, uint32_t *count);
410 static void delete_unused_nodes(di_devlink_handle_t hdp, cache_node_t *cnp);
H A Dinstance.h130 void e_ddi_borrow_instance(dev_info_t *cdip, in_node_t *cnp);
131 void e_ddi_return_instance(dev_info_t *cdip, char *addr, in_node_t *cnp);
H A Dses_node.c138 ses_node_t *np, *pnp, *cnp, *root; local
299 cnp = ses_node_alloc(sp, np);
300 if (cnp == NULL)
303 cnp->sn_type = SES_NODE_ELEMENT;
304 SES_NV_ADD(uint64, err, cnp->sn_props,
306 SES_NV_ADD(uint64, err, cnp->sn_props,
308 SES_NV_ADD(uint64, err, cnp->sn_props,
310 SES_NV_ADD(uint64, err, cnp->sn_props,
H A Ddt_pragma.c233 dt_pragma_depends(const char *prname, dt_node_t *cnp) argument
236 dt_node_t *nnp = cnp ? cnp->dn_list : NULL;
241 if (cnp == NULL || nnp == NULL ||
242 cnp->dn_kind != DT_NODE_IDENT || nnp->dn_kind != DT_NODE_IDENT) {
247 if (strcmp(cnp->dn_string, "provider") == 0)
249 else if (strcmp(cnp->dn_string, "module") == 0) {
252 } else if (strcmp(cnp->dn_string, "library") == 0) {
299 "specified by #pragma %s\n", cnp->dn_string, prname);
304 cnp
H A Ddt_cc.c1563 dt_compile_one_clause(dtrace_hdl_t *dtp, dt_node_t *cnp, dt_node_t *pnp) argument
1571 (void) dt_node_cook(cnp, DT_IDFLG_REF);
1574 dt_node_printr(cnp, stderr, 0);
1582 if (cnp->dn_pred != NULL) {
1583 dt_cg(yypcb, cnp->dn_pred);
1587 if (cnp->dn_acts == NULL) {
1589 cnp->dn_ctxattr, _dtrace_defattr), cnp);
1592 for (dnp = cnp->dn_acts; dnp != NULL; dnp = dnp->dn_list) {
1594 sdp = dt_stmt_create(dtp, edp, cnp
1625 dt_compile_clause(dtrace_hdl_t *dtp, dt_node_t *cnp) argument
H A Dpromif_stree.c332 prom_node_t *cnp; local
344 cnp = create_node(pnp, child);
345 pnp->pn_child = cnp;
346 create_peers(cnp, child);
347 pnp = cnp;
H A Dgenassym.c137 ga_parse_name(char *line, char **cnp, char **szdp, char **shdp) argument
139 return (ga_parse_common(line, 1, 3, cnp, szdp, shdp));
H A Dinstance.c1591 e_ddi_borrow_instance(dev_info_t *cdip, in_node_t *cnp) argument
1604 if (cnp->in_drivers) {
1607 cnp->in_drivers->ind_driver_name);
1635 cnp->in_drivers = anp->in_drivers;
1643 e_ddi_return_instance(dev_info_t *cdip, char *addr, in_node_t *cnp) argument
1660 if (cnp->in_drivers == NULL) {
1661 ddi_err(DER_PANIC, cdip, "cnp has no inst: %p", cnp);
1670 anp->in_drivers = cnp->in_drivers;
1671 cnp
H A Decore.tex1285 When configuring the device to act as notification point, the ecore will initialize the NIG block accordingly and pass the priority vlan and cnp send timeout values to FW. When configuring the device to act as reaction point, the ecore will send a ramrod to FW that configures the rate limiters allocated for dcqcn support with the values received from the upper layer driver ( such as maximum rate, byte counter limit, active increase rate etc... full detail in ecore\_roce\_api.h file ). At this point all rate limiters will be configured with the same values. If in the future there will be a need to configure different rate limiters with different values an additional API function will be provided. During initialization, ecore will map between physical queues used for RoCE and rate limiters. The number of rate limiters allocated is handled by resource management and is currently divided equally between the functions. During modify\_qp, ecore will configure the responder and requester to work with a unique physical queue, which is configured to work with a unique rate limiter. QPs that are opened after rate limiters are used out will be configured to run on a default physical queue which does not have a rate limiter. FW assumes that the qp\_id is equal to the physical queue id. For simplicity, the implementation assumes that Ethernet is not run simultaneously with RoCE (i.e. Roce only personality). If dcqcn is enabled and ethernet is run, ethernet will run on the same physical queue as the first qp that is allocated.

Completed in 79 milliseconds