/Linux-v5.4/drivers/net/ethernet/netronome/nfp/crypto/ |
D | tls.c | 188 return &req->back; in nfp_net_tls_set_ipv4() 208 return &req->back; in nfp_net_tls_set_ipv6() 213 struct nfp_crypto_req_add_back *back, struct sock *sk, in nfp_net_tls_set_l4() argument 221 back->src_port = 0; in nfp_net_tls_set_l4() 222 back->dst_port = 0; in nfp_net_tls_set_l4() 224 back->src_port = inet->inet_dport; in nfp_net_tls_set_l4() 225 back->dst_port = inet->inet_sport; in nfp_net_tls_set_l4() 272 struct nfp_crypto_req_add_back *back; in nfp_net_tls_add() local 327 back = nfp_net_tls_set_ipv6(nn, req, sk, direction); in nfp_net_tls_add() 329 back = nfp_net_tls_set_ipv4(nn, req, sk, direction); in nfp_net_tls_add() [all …]
|
D | fw.h | 53 struct nfp_crypto_req_add_back back; member 60 struct nfp_crypto_req_add_back back; member
|
/Linux-v5.4/Documentation/scsi/ |
D | sd-parameters.txt | 12 write back | 1 0 | on | on 13 write back, no read (daft) | 1 1 | on | off 15 To set cache type to "write back" and save this setting to the drive: 17 # echo "write back" > cache_type 22 # echo "temporary write back" > cache_type
|
/Linux-v5.4/fs/erofs/ |
D | namei.c | 51 int head, back; in find_target_dirent() local 57 back = ndirents - 1; in find_target_dirent() 60 while (head <= back) { in find_target_dirent() 61 const int mid = head + (back - head) / 2; in find_target_dirent() 82 back = mid - 1; in find_target_dirent() 95 int head, back; in find_target_block_classic() local 101 back = erofs_inode_datablocks(dir) - 1; in find_target_block_classic() 103 while (head <= back) { in find_target_block_classic() 104 const int mid = head + (back - head) / 2; in find_target_block_classic() 155 back = mid - 1; in find_target_block_classic()
|
/Linux-v5.4/lib/lz4/ |
D | lz4hc_compress.c | 208 int back = 0; in LZ4HC_InsertAndGetWiderMatch() local 210 while ((ip + back > iLowLimit) in LZ4HC_InsertAndGetWiderMatch() 211 && (matchPtr + back > lowPrefixPtr) in LZ4HC_InsertAndGetWiderMatch() 212 && (ip[back - 1] == matchPtr[back - 1])) in LZ4HC_InsertAndGetWiderMatch() 213 back--; in LZ4HC_InsertAndGetWiderMatch() 215 mlt -= back; in LZ4HC_InsertAndGetWiderMatch() 219 *matchpos = matchPtr + back; in LZ4HC_InsertAndGetWiderMatch() 220 *startpos = ip + back; in LZ4HC_InsertAndGetWiderMatch() 229 int back = 0; in LZ4HC_InsertAndGetWiderMatch() local 241 while ((ip + back > iLowLimit) in LZ4HC_InsertAndGetWiderMatch() [all …]
|
/Linux-v5.4/drivers/net/ethernet/intel/ice/ |
D | ice_lib.c | 17 struct ice_hw *hw = &vsi->back->hw; in ice_setup_rx_ctx() 90 dev_err(&vsi->back->pdev->dev, in ice_setup_rx_ctx() 119 struct ice_hw *hw = &vsi->back->hw; in ice_setup_tx_ctx() 205 struct ice_pf *pf = vsi->back; in ice_vsi_ctrl_rx_ring() 260 struct ice_pf *pf = vsi->back; in ice_vsi_alloc_arrays() 322 dev_dbg(&vsi->back->pdev->dev, in ice_vsi_set_num_desc() 338 struct ice_pf *pf = vsi->back; in ice_vsi_set_num_qs() 420 struct ice_pf *pf = vsi->back; in ice_vsi_delete() 448 struct ice_pf *pf = vsi->back; in ice_vsi_free_arrays() 489 if (!vsi->back) in ice_vsi_clear() [all …]
|
D | ice_ethtool.c | 168 struct ice_pf *pf = vsi->back; in ice_get_drvinfo() 188 struct ice_pf *pf = np->vsi->back; in ice_get_regs() 202 struct ice_pf *pf = np->vsi->back; in ice_get_msglevel() 216 struct ice_pf *pf = np->vsi->back; in ice_set_msglevel() 231 struct ice_pf *pf = np->vsi->back; in ice_get_eeprom_len() 243 struct ice_pf *pf = vsi->back; in ice_get_eeprom() 330 struct ice_pf *pf = np->vsi->back; in ice_eeprom_test() 344 struct ice_pf *pf = (struct ice_pf *)hw->back; in ice_reg_pattern_test() 646 struct ice_pf *pf = orig_vsi->back; in ice_loopback_test() 745 struct ice_pf *pf = np->vsi->back; in ice_intr_test() [all …]
|
/Linux-v5.4/Documentation/driver-api/md/ |
D | raid5-cache.rst | 8 since 4.4) or write-back mode (supported since 4.10). mdadm (supported since 11 in write-through mode. A user can switch it to write-back mode by:: 13 echo "write-back" > /sys/block/md0/md/journal_mode 15 And switch it back to write-through mode by:: 30 tries to resync the array to bring it back to normal state. But before the 47 write-back mode 50 write-back mode fixes the 'write hole' issue too, since all write data is 51 cached on cache disk. But the main goal of 'write-back' cache is to speed up 56 overhead too. Write-back cache will aggregate the data and flush the data to 61 In write-back mode, MD reports IO completion to upper layer (usually [all …]
|
/Linux-v5.4/fs/xfs/libxfs/ |
D | xfs_da_format.c | 438 to->back = be32_to_cpu(from->hdr.info.back); in xfs_dir2_leaf_hdr_from_disk() 456 to->hdr.info.back = cpu_to_be32(from->back); in xfs_dir2_leaf_hdr_to_disk() 470 to->back = be32_to_cpu(hdr3->info.hdr.back); in xfs_dir3_leaf_hdr_from_disk() 490 hdr3->info.hdr.back = cpu_to_be32(from->back); in xfs_dir3_leaf_hdr_to_disk() 519 to->back = be32_to_cpu(from->hdr.info.back); in xfs_da2_node_hdr_from_disk() 532 to->hdr.info.back = cpu_to_be32(from->back); in xfs_da2_node_hdr_to_disk() 547 to->back = be32_to_cpu(hdr3->info.hdr.back); in xfs_da3_node_hdr_from_disk() 562 hdr3->info.hdr.back = cpu_to_be32(from->back); in xfs_da3_node_hdr_to_disk()
|
D | xfs_da_btree.c | 511 node->hdr.info.back = cpu_to_be32(oldblk->blkno); in xfs_da3_split() 517 if (node->hdr.info.back) { in xfs_da3_split() 518 if (be32_to_cpu(node->hdr.info.back) != addblk->blkno) { in xfs_da3_split() 1065 ASSERT(!blkinfo->back); in xfs_da_blkinfo_onlychild_validate() 1097 ASSERT(oldroothdr.back == 0); in xfs_da3_root_join() 1218 forward = nodehdr.forw < nodehdr.back; in xfs_da3_node_toosmall() 1224 blkno = nodehdr.back; in xfs_da3_node_toosmall() 1745 new_info->back = old_info->back; in xfs_da3_blk_link() 1746 if (old_info->back) { in xfs_da3_blk_link() 1748 be32_to_cpu(old_info->back), in xfs_da3_blk_link() [all …]
|
/Linux-v5.4/scripts/coccinelle/iterators/ |
D | list_entry_update.cocci | 5 /// that there is a path from the reassignment back to the top of the loop. 39 @back depends on (org || report) && !context exists@ 49 @script:python depends on back && org@ 57 @script:python depends on back && report@
|
/Linux-v5.4/drivers/md/bcache/ |
D | util.h | 118 size_t front, back, size, mask; \ 124 c = (fifo)->data[iter], iter != (fifo)->back; \ 136 (fifo)->front = (fifo)->back = 0; \ 162 #define fifo_used(fifo) (((fifo)->back - (fifo)->front) & (fifo)->mask) 170 ((fifo)->data[((fifo)->back - 1) & (fifo)->mask]) 178 (fifo)->data[(fifo)->back++] = (i); \ 179 (fifo)->back &= (fifo)->mask; \ 209 --(fifo)->back; \ 210 (fifo)->back &= (fifo)->mask; \ 211 (i) = (fifo)->data[(fifo)->back] \ [all …]
|
/Linux-v5.4/drivers/net/ethernet/intel/iavf/ |
D | iavf_client.c | 42 params->link_up = vsi->back->link_up; in iavf_client_get_params() 65 cinst = vsi->back->cinst; in iavf_notify_client_message() 68 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_message() 90 cinst = vsi->back->cinst; in iavf_notify_client_l2_params() 94 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_l2_params() 112 struct iavf_adapter *adapter = vsi->back; in iavf_notify_client_open() 118 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_open() 164 struct iavf_adapter *adapter = vsi->back; in iavf_notify_client_close() 169 dev_dbg(&vsi->back->pdev->dev, in iavf_notify_client_close()
|
/Linux-v5.4/Documentation/ABI/testing/ |
D | sysfs-class-bdi | 36 total write-back cache that relates to its current average 40 percentage of the write-back cache to a particular device. 46 given percentage of the write-back cache. This is useful in 48 most of the write-back cache. For example in case of an NFS
|
D | sysfs-platform-hidma-mgmt | 63 read transactions that can be issued back to back. 84 write transactions that can be issued back to back.
|
/Linux-v5.4/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
D | hclge_mbx.c | 21 struct hclge_dev *hdev = vport->back; in hclge_gen_resp_to_vf() 63 struct hclge_dev *hdev = vport->back; in hclge_send_mbx_msg() 88 struct hclge_dev *hdev = vport->back; in hclge_inform_reset_assert_to_vf() 213 return hclge_cmd_set_promisc_mode(vport->back, ¶m); in hclge_set_vf_promisc_mode() 220 struct hclge_dev *hdev = vport->back; in hclge_set_vf_uc_mac_addr() 264 struct hclge_dev *hdev = vport->back; in hclge_set_vf_mc_mac_addr() 390 struct hclge_dev *hdev = vport->back; in hclge_get_vf_queue_info() 407 struct hclge_dev *hdev = vport->back; in hclge_get_vf_queue_depth() 419 struct hclge_dev *hdev = vport->back; in hclge_get_vf_media_type() 431 struct hclge_dev *hdev = vport->back; in hclge_get_link_info() [all …]
|
D | hclge_dcb.c | 68 struct hclge_dev *hdev = vport->back; in hclge_ieee_getets() 227 struct hclge_dev *hdev = vport->back; in hclge_ieee_setets() 283 struct hclge_dev *hdev = vport->back; in hclge_ieee_getpfc() 319 struct hclge_dev *hdev = vport->back; in hclge_ieee_setpfc() 372 struct hclge_dev *hdev = vport->back; in hclge_getdcbx() 384 struct hclge_dev *hdev = vport->back; in hclge_setdcbx() 403 struct hclge_dev *hdev = vport->back; in hclge_setup_tc()
|
/Linux-v5.4/drivers/net/ethernet/intel/i40e/ |
D | i40e_main.c | 118 struct i40e_pf *pf = (struct i40e_pf *)hw->back; in i40e_allocate_dma_mem_d() 136 struct i40e_pf *pf = (struct i40e_pf *)hw->back; in i40e_free_dma_mem_d() 308 struct i40e_pf *pf = vsi->back; in i40e_tx_timeout() 626 struct i40e_pf *pf = vsi->back; in i40e_update_eth_stats() 772 struct i40e_pf *pf = vsi->back; in i40e_update_vsi_stats() 1104 struct i40e_pf *pf = vsi->back; in i40e_update_stats() 1318 struct i40e_pf *pf = vsi->back; in i40e_rm_default_mac_filter() 1381 set_bit(__I40E_MACVLAN_SYNC_PENDING, vsi->back->state); in i40e_add_filter() 1431 set_bit(__I40E_MACVLAN_SYNC_PENDING, vsi->back->state); in __i40e_del_filter() 1536 struct i40e_pf *pf = vsi->back; in i40e_set_mac() [all …]
|
D | i40e_client.c | 49 struct i40e_dcbx_config *dcb_cfg = &vsi->back->hw.local_dcbx_config; in i40e_client_get_params() 64 dev_err(&vsi->back->pdev->dev, "Invalid queue set handle for TC = %d, vsi id = %d\n", in i40e_client_get_params() 86 struct i40e_pf *pf = vsi->back; in i40e_notify_client_of_vf_msg() 112 struct i40e_pf *pf = vsi->back; in i40e_notify_client_of_l2_param_changes() 119 dev_dbg(&vsi->back->pdev->dev, in i40e_notify_client_of_l2_param_changes() 124 dev_dbg(&vsi->back->pdev->dev, "Client is not open, abort l2 param change\n"); in i40e_notify_client_of_l2_param_changes() 171 struct i40e_pf *pf = vsi->back; in i40e_notify_client_of_netdev_close() 177 dev_dbg(&vsi->back->pdev->dev, in i40e_notify_client_of_netdev_close() 724 err = i40e_aq_update_vsi_params(&vsi->back->hw, &ctxt, NULL); in i40e_client_update_vsi_ctxt()
|
D | i40e_ethtool.c | 1048 struct i40e_pf *pf = np->vsi->back; in i40e_get_link_ksettings() 1139 struct i40e_pf *pf = np->vsi->back; in i40e_set_link_ksettings() 1375 struct i40e_pf *pf = np->vsi->back; in i40e_set_fec_cfg() 1436 struct i40e_pf *pf = np->vsi->back; in i40e_get_fec_param() 1479 struct i40e_pf *pf = np->vsi->back; in i40e_set_fec_param() 1523 struct i40e_pf *pf = np->vsi->back; in i40e_nway_reset() 1550 struct i40e_pf *pf = np->vsi->back; in i40e_get_pauseparam() 1585 struct i40e_pf *pf = np->vsi->back; in i40e_set_pauseparam() 1678 struct i40e_pf *pf = np->vsi->back; in i40e_get_msglevel() 1690 struct i40e_pf *pf = np->vsi->back; in i40e_set_msglevel() [all …]
|
/Linux-v5.4/arch/mips/include/asm/octeon/ |
D | cvmx-packet.h | 52 uint64_t back:4; member 63 uint64_t back:4;
|
D | cvmx-helper-util.h | 146 start_of_buffer = ((buffer_ptr.s.addr >> 7) - buffer_ptr.s.back) << 7; in cvmx_helper_free_packet_data() 160 ((buffer_ptr.s.addr >> 7) - buffer_ptr.s.back) << 7; in cvmx_helper_free_packet_data()
|
/Linux-v5.4/Documentation/driver-api/ |
D | dell_rbu.rst | 36 maintains a link list of packets for reading them back. 74 packets of data arranged back to back. It can be done as follows 79 image file and then arrange all these packets back to back in to one single 120 read back the image downloaded.
|
/Linux-v5.4/drivers/gpu/drm/udl/ |
D | udl_transfer.c | 39 const unsigned long *back = (const unsigned long *) bback; 47 if (back[j] != front[j]) { 54 if (back[k] != front[k]) {
|
/Linux-v5.4/Documentation/devicetree/bindings/dma/ |
D | qcom_hidma_mgmt.txt | 32 applied back to back while writing to the destination before yielding 35 applied back to back while reading the source before yielding the bus.
|