Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

dmaengine: Fix spelling mistakes

Correct spelling mistakes in the DMA engine to improve readability
and clarity without altering functionality.

Signed-off-by: Amit Vadhavana <av2082000@gmail.com>
Reviewed-by: Kees Cook <kees@kernel.org>
Reviewed-by: Dave Jiang <dave.jiang@intel.com>
Reviewed-by: Fenghua Yu <fenghua.yu@intel.com>
Link: https://lore.kernel.org/r/20240831172949.13189-1-av2082000@gmail.com
Signed-off-by: Vinod Koul <vkoul@kernel.org>

authored by

Amit Vadhavana and committed by
Vinod Koul
a688efea 5d318b59

+49 -49
+2 -2
drivers/dma/acpi-dma.c
··· 112 112 } 113 113 114 114 /** 115 - * acpi_dma_parse_csrt - parse CSRT to exctract additional DMA resources 115 + * acpi_dma_parse_csrt - parse CSRT to extract additional DMA resources 116 116 * @adev: ACPI device to match with 117 117 * @adma: struct acpi_dma of the given DMA controller 118 118 * ··· 305 305 * found. 306 306 * 307 307 * Return: 308 - * 0, if no information is avaiable, -1 on mismatch, and 1 otherwise. 308 + * 0, if no information is available, -1 on mismatch, and 1 otherwise. 309 309 */ 310 310 static int acpi_dma_update_dma_spec(struct acpi_dma *adma, 311 311 struct acpi_dma_spec *dma_spec)
+2 -2
drivers/dma/altera-msgdma.c
··· 153 153 /** 154 154 * struct msgdma_sw_desc - implements a sw descriptor 155 155 * @async_tx: support for the async_tx api 156 - * @hw_desc: assosiated HW descriptor 156 + * @hw_desc: associated HW descriptor 157 157 * @node: node to move from the free list to the tx list 158 158 * @tx_list: transmit list node 159 159 */ ··· 511 511 * of the DMA controller. The descriptor will get flushed to the 512 512 * FIFO, once the last word (control word) is written. Since we 513 513 * are not 100% sure that memcpy() writes all word in the "correct" 514 - * oder (address from low to high) on all architectures, we make 514 + * order (address from low to high) on all architectures, we make 515 515 * sure this control word is written last by single coding it and 516 516 * adding some write-barriers here. 517 517 */
+1 -1
drivers/dma/amba-pl08x.c
··· 2 2 /* 3 3 * Copyright (c) 2006 ARM Ltd. 4 4 * Copyright (c) 2010 ST-Ericsson SA 5 - * Copyirght (c) 2017 Linaro Ltd. 5 + * Copyright (c) 2017 Linaro Ltd. 6 6 * 7 7 * Author: Peter Pearse <peter.pearse@arm.com> 8 8 * Author: Linus Walleij <linus.walleij@linaro.org>
+3 -3
drivers/dma/at_hdmac.c
··· 339 339 * @regs: memory mapped register base 340 340 * @clk: dma controller clock 341 341 * @save_imr: interrupt mask register that is saved on suspend/resume cycle 342 - * @all_chan_mask: all channels availlable in a mask 342 + * @all_chan_mask: all channels available in a mask 343 343 * @lli_pool: hw lli table 344 344 * @memset_pool: hw memset pool 345 345 * @chan: channels table to store at_dma_chan structures ··· 668 668 * CTRLA is read in turn, next the DSCR is read a second time. If the two 669 669 * consecutive read values of the DSCR are the same then we assume both refers 670 670 * to the very same LLI as well as the CTRLA value read inbetween does. For 671 - * cyclic tranfers, the assumption is that a full loop is "not so fast". If the 671 + * cyclic transfers, the assumption is that a full loop is "not so fast". If the 672 672 * two DSCR values are different, we read again the CTRLA then the DSCR till two 673 673 * consecutive read values from DSCR are equal or till the maximum trials is 674 674 * reach. This algorithm is very unlikely not to find a stable value for DSCR. ··· 700 700 break; 701 701 702 702 /* 703 - * DSCR has changed inside the DMA controller, so the previouly 703 + * DSCR has changed inside the DMA controller, so the previously 704 704 * read value of CTRLA may refer to an already processed 705 705 * descriptor hence could be outdated. We need to update ctrla 706 706 * to match the current descriptor.
+2 -2
drivers/dma/bcm-sba-raid.c
··· 15 15 * number of hardware rings over one or more SBA hardware devices. By 16 16 * design, the internal buffer size of SBA hardware device is limited 17 17 * but all offload operations supported by SBA can be broken down into 18 - * multiple small size requests and executed parallely on multiple SBA 18 + * multiple small size requests and executed parallelly on multiple SBA 19 19 * hardware devices for achieving high through-put. 20 20 * 21 21 * The Broadcom SBA RAID driver does not require any register programming ··· 135 135 u32 max_xor_srcs; 136 136 u32 max_resp_pool_size; 137 137 u32 max_cmds_pool_size; 138 - /* Maibox client and Mailbox channels */ 138 + /* Mailbox client and Mailbox channels */ 139 139 struct mbox_client client; 140 140 struct mbox_chan *mchan; 141 141 struct device *mbox_dev;
+1 -1
drivers/dma/bcm2835-dma.c
··· 369 369 /* the last frame requires extra flags */ 370 370 d->cb_list[d->frames - 1].cb->info |= finalextrainfo; 371 371 372 - /* detect a size missmatch */ 372 + /* detect a size mismatch */ 373 373 if (buf_len && (d->size != buf_len)) 374 374 goto error_cb; 375 375
+2 -2
drivers/dma/ep93xx_dma.c
··· 841 841 desc = container_of(tx, struct ep93xx_dma_desc, txd); 842 842 843 843 /* 844 - * If nothing is currently prosessed, we push this descriptor 844 + * If nothing is currently processed, we push this descriptor 845 845 * directly to the hardware. Otherwise we put the descriptor 846 846 * to the pending queue. 847 847 */ ··· 1025 1025 * @chan: channel 1026 1026 * @sgl: list of buffers to transfer 1027 1027 * @sg_len: number of entries in @sgl 1028 - * @dir: direction of tha DMA transfer 1028 + * @dir: direction of the DMA transfer 1029 1029 * @flags: flags for the descriptor 1030 1030 * @context: operation context (ignored) 1031 1031 *
+3 -3
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.h
··· 12 12 u32 rsv:32; 13 13 union { 14 14 struct { 15 - u32 ssd:12; /* souce stride distance */ 16 - u32 sss:12; /* souce stride size */ 15 + u32 ssd:12; /* source stride distance */ 16 + u32 sss:12; /* source stride size */ 17 17 u32 rsv1:8; 18 18 } sdf; 19 19 struct { ··· 48 48 #define QDMA_SER_DISABLE (8) /* no notification */ 49 49 #define QDMA_SER_CTX BIT(8) /* notification by FQD_CTX[fqid] */ 50 50 #define QDMA_SER_DEST (2 << 8) /* notification by destination desc */ 51 - #define QDMA_SER_BOTH (3 << 8) /* soruce and dest notification */ 51 + #define QDMA_SER_BOTH (3 << 8) /* source and dest notification */ 52 52 #define QDMA_FD_SPF_ENALBE BIT(30) /* source prefetch enable */ 53 53 54 54 #define QMAN_FD_VA_ENABLE BIT(14) /* Address used is virtual address */
+1 -1
drivers/dma/hisi_dma.c
··· 677 677 writel_relaxed(tmp, addr); 678 678 679 679 /* 680 - * 0 - dma should process FLR whith CPU. 680 + * 0 - dma should process FLR with CPU. 681 681 * 1 - dma not process FLR, only cpu process FLR. 682 682 */ 683 683 addr = q_base + HISI_DMA_HIP09_DMA_FLR_DISABLE +
+2 -2
drivers/dma/idma64.c
··· 290 290 desc->length += hw->len; 291 291 } while (i); 292 292 293 - /* Trigger an interrupt after the last block is transfered */ 293 + /* Trigger an interrupt after the last block is transferred */ 294 294 lli->ctllo |= IDMA64C_CTLL_INT_EN; 295 295 296 296 /* Disable LLP transfer in the last block */ ··· 364 364 if (!i) 365 365 return bytes; 366 366 367 - /* The current chunk is not fully transfered yet */ 367 + /* The current chunk is not fully transferred yet */ 368 368 bytes += desc->hw[--i].len; 369 369 370 370 return bytes - IDMA64C_CTLH_BLOCK_TS(ctlhi);
+1 -1
drivers/dma/idxd/submit.c
··· 134 134 * completing the descriptor will return desc to allocator and 135 135 * the desc can be acquired by a different process and the 136 136 * desc->list can be modified. Delete desc from list so the 137 - * list trasversing does not get corrupted by the other process. 137 + * list traversing does not get corrupted by the other process. 138 138 */ 139 139 list_for_each_entry_safe(d, t, &flist, list) { 140 140 list_del_init(&d->list);
+1 -1
drivers/dma/ioat/init.c
··· 905 905 906 906 op = IOAT_OP_XOR_VAL; 907 907 908 - /* validate the sources with the destintation page */ 908 + /* validate the sources with the destination page */ 909 909 for (i = 0; i < IOAT_NUM_SRC_TEST; i++) 910 910 xor_val_srcs[i] = xor_srcs[i]; 911 911 xor_val_srcs[i] = dest;
+1 -1
drivers/dma/lgm/lgm-dma.c
··· 107 107 * If header mode is set in DMA descriptor, 108 108 * If bit 30 is disabled, HDR_LEN must be configured according to channel 109 109 * requirement. 110 - * If bit 30 is enabled(checksum with heade mode), HDR_LEN has no need to 110 + * If bit 30 is enabled(checksum with header mode), HDR_LEN has no need to 111 111 * be configured. It will enable check sum for switch 112 112 * If header mode is not set in DMA descriptor, 113 113 * This register setting doesn't matter
+2 -2
drivers/dma/ls2x-apb-dma.c
··· 33 33 #define LDMA_STOP BIT(4) /* DMA stop operation */ 34 34 #define LDMA_CONFIG_MASK GENMASK(4, 0) /* DMA controller config bits mask */ 35 35 36 - /* Bitfields in ndesc_addr field of HW decriptor */ 36 + /* Bitfields in ndesc_addr field of HW descriptor */ 37 37 #define LDMA_DESC_EN BIT(0) /*1: The next descriptor is valid */ 38 38 #define LDMA_DESC_ADDR_LOW GENMASK(31, 1) 39 39 40 - /* Bitfields in cmd field of HW decriptor */ 40 + /* Bitfields in cmd field of HW descriptor */ 41 41 #define LDMA_INT BIT(1) /* Enable DMA interrupts */ 42 42 #define LDMA_DATA_DIRECTION BIT(12) /* 1: write to device, 0: read from device */ 43 43
+2 -2
drivers/dma/mediatek/mtk-cqdma.c
··· 518 518 /* setup dma channel */ 519 519 cvd[i]->ch = c; 520 520 521 - /* setup sourece, destination, and length */ 521 + /* setup source, destination, and length */ 522 522 tlen = (len > MTK_CQDMA_MAX_LEN) ? MTK_CQDMA_MAX_LEN : len; 523 523 cvd[i]->len = tlen; 524 524 cvd[i]->src = src; ··· 617 617 u32 i, min_refcnt = U32_MAX, refcnt; 618 618 unsigned long flags; 619 619 620 - /* allocate PC with the minimun refcount */ 620 + /* allocate PC with the minimum refcount */ 621 621 for (i = 0; i < cqdma->dma_channels; ++i) { 622 622 refcnt = refcount_read(&cqdma->pc[i]->refcnt); 623 623 if (refcnt < min_refcnt) {
+1 -1
drivers/dma/mediatek/mtk-hsdma.c
··· 226 226 * @pc_refcnt: Track how many VCs are using the PC 227 227 * @lock: Lock protect agaisting multiple VCs access PC 228 228 * @soc: The pointer to area holding differences among 229 - * vaious platform 229 + * various platform 230 230 */ 231 231 struct mtk_hsdma_device { 232 232 struct dma_device ddev;
+2 -2
drivers/dma/mv_xor.c
··· 414 414 if (!mv_chan_is_busy(mv_chan)) { 415 415 u32 current_desc = mv_chan_get_current_desc(mv_chan); 416 416 /* 417 - * and the curren desc is the end of the chain before 417 + * and the current desc is the end of the chain before 418 418 * the append, then we need to start the channel 419 419 */ 420 420 if (current_desc == old_chain_tail->async_tx.phys) ··· 1074 1074 if (!mv_chan->dma_desc_pool_virt) 1075 1075 return ERR_PTR(-ENOMEM); 1076 1076 1077 - /* discover transaction capabilites from the platform data */ 1077 + /* discover transaction capabilities from the platform data */ 1078 1078 dma_dev->cap_mask = cap_mask; 1079 1079 1080 1080 INIT_LIST_HEAD(&dma_dev->channels);
+1 -1
drivers/dma/mv_xor.h
··· 99 99 * @common: common dmaengine channel object members 100 100 * @slots_allocated: records the actual size of the descriptor slot pool 101 101 * @irq_tasklet: bottom half where mv_xor_slot_cleanup runs 102 - * @op_in_desc: new mode of driver, each op is writen to descriptor. 102 + * @op_in_desc: new mode of driver, each op is written to descriptor. 103 103 */ 104 104 struct mv_xor_chan { 105 105 int pending;
+1 -1
drivers/dma/mv_xor_v2.c
··· 175 175 * struct mv_xor_v2_sw_desc - implements a xor SW descriptor 176 176 * @idx: descriptor index 177 177 * @async_tx: support for the async_tx api 178 - * @hw_desc: assosiated HW descriptor 178 + * @hw_desc: associated HW descriptor 179 179 * @free_list: node of the free SW descriprots list 180 180 */ 181 181 struct mv_xor_v2_sw_desc {
+1 -1
drivers/dma/nbpfaxi.c
··· 897 897 /* 898 898 * We could check config->slave_id to match chan->terminal here, 899 899 * but with DT they would be coming from the same source, so 900 - * such a check would be superflous 900 + * such a check would be superfluous 901 901 */ 902 902 903 903 chan->slave_dst_addr = config->dst_addr;
+2 -2
drivers/dma/of-dma.c
··· 26 26 * 27 27 * Finds a DMA controller with matching device node and number for dma cells 28 28 * in a list of registered DMA controllers. If a match is found a valid pointer 29 - * to the DMA data stored is retuned. A NULL pointer is returned if no match is 29 + * to the DMA data stored is returned. A NULL pointer is returned if no match is 30 30 * found. 31 31 */ 32 32 static struct of_dma *of_dma_find_controller(const struct of_phandle_args *dma_spec) ··· 342 342 * 343 343 * This function can be used as the of xlate callback for DMA driver which wants 344 344 * to match the channel based on the channel id. When using this xlate function 345 - * the #dma-cells propety of the DMA controller dt node needs to be set to 1. 345 + * the #dma-cells property of the DMA controller dt node needs to be set to 1. 346 346 * The data parameter of of_dma_controller_register must be a pointer to the 347 347 * dma_device struct the function should match upon. 348 348 *
+1 -1
drivers/dma/owl-dma.c
··· 1156 1156 } 1157 1157 1158 1158 /* 1159 - * Eventhough the DMA controller is capable of generating 4 1159 + * Even though the DMA controller is capable of generating 4 1160 1160 * IRQ's for DMA priority feature, we only use 1 IRQ for 1161 1161 * simplification. 1162 1162 */
+1 -1
drivers/dma/ppc4xx/adma.c
··· 9 9 */ 10 10 11 11 /* 12 - * This driver supports the asynchrounous DMA copy and RAID engines available 12 + * This driver supports the asynchronous DMA copy and RAID engines available 13 13 * on the AMCC PPC440SPe Processors. 14 14 * Based on the Intel Xscale(R) family of I/O Processors (IOP 32x, 33x, 134x) 15 15 * ADMA driver written by D.Williams.
+1 -1
drivers/dma/ppc4xx/dma.h
··· 14 14 15 15 /* Number of elements in the array with statical CDBs */ 16 16 #define MAX_STAT_DMA_CDBS 16 17 - /* Number of DMA engines available on the contoller */ 17 + /* Number of DMA engines available on the controller */ 18 18 #define DMA_ENGINES_NUM 2 19 19 20 20 /* Maximum h/w supported number of destinations */
+1 -1
drivers/dma/ptdma/ptdma.h
··· 192 192 /* Queue dma pool */ 193 193 struct dma_pool *dma_pool; 194 194 195 - /* Queue base address (not neccessarily aligned)*/ 195 + /* Queue base address (not necessarily aligned)*/ 196 196 struct ptdma_desc *qbase; 197 197 198 198 /* Aligned queue start address (per requirement) */
+2 -2
drivers/dma/qcom/bam_dma.c
··· 440 440 val |= BAM_EN; 441 441 writel_relaxed(val, bam_addr(bdev, 0, BAM_CTRL)); 442 442 443 - /* set descriptor threshhold, start with 4 bytes */ 443 + /* set descriptor threshold, start with 4 bytes */ 444 444 writel_relaxed(DEFAULT_CNT_THRSHLD, 445 445 bam_addr(bdev, 0, BAM_DESC_CNT_TRSHLD)); 446 446 ··· 667 667 for_each_sg(sgl, sg, sg_len, i) 668 668 num_alloc += DIV_ROUND_UP(sg_dma_len(sg), BAM_FIFO_SIZE); 669 669 670 - /* allocate enough room to accomodate the number of entries */ 670 + /* allocate enough room to accommodate the number of entries */ 671 671 async_desc = kzalloc(struct_size(async_desc, desc, num_alloc), 672 672 GFP_NOWAIT); 673 673
+1 -1
drivers/dma/qcom/gpi.c
··· 1856 1856 1857 1857 read_lock_irqsave(&gpii->pm_lock, pm_lock_flags); 1858 1858 1859 - /* move all submitted discriptors to issued list */ 1859 + /* move all submitted descriptors to issued list */ 1860 1860 spin_lock_irqsave(&gchan->vc.lock, flags); 1861 1861 if (vchan_issue_pending(&gchan->vc)) 1862 1862 vd = list_last_entry(&gchan->vc.desc_issued,
+1 -1
drivers/dma/qcom/qcom_adm.c
··· 650 650 /* 651 651 * residue is either the full length if it is in the issued list, or 0 652 652 * if it is in progress. We have no reliable way of determining 653 - * anything inbetween 653 + * anything in between 654 654 */ 655 655 dma_set_residue(txstate, residue); 656 656
+1 -1
drivers/dma/sh/shdmac.c
··· 318 318 } 319 319 320 320 /* 321 - * Find a slave channel configuration from the contoller list by either a slave 321 + * Find a slave channel configuration from the controller list by either a slave 322 322 * ID in the non-DT case, or by a MID/RID value in the DT case 323 323 */ 324 324 static const struct sh_dmae_slave_config *dmae_find_slave(
+1 -1
drivers/dma/ste_dma40.h
··· 4 4 #define STE_DMA40_H 5 5 6 6 /* 7 - * Maxium size for a single dma descriptor 7 + * Maximum size for a single dma descriptor 8 8 * Size is limited to 16 bits. 9 9 * Size is in the units of addr-widths (1,2,4,8 bytes) 10 10 * Larger transfers will be split up to multiple linked desc
+1 -1
drivers/dma/ste_dma40_ll.h
··· 369 369 * @lcsp02: Either maps to register lcsp0 if src or lcsp2 if dst. 370 370 * @lcsp13: Either maps to register lcsp1 if src or lcsp3 if dst. 371 371 * 372 - * This struct must be 8 bytes aligned since it will be accessed directy by 372 + * This struct must be 8 bytes aligned since it will be accessed directly by 373 373 * the DMA. Never add any none hw mapped registers to this struct. 374 374 */ 375 375
+1 -1
drivers/dma/tegra20-apb-dma.c
··· 463 463 464 464 /* 465 465 * If interrupt is pending then do nothing as the ISR will handle 466 - * the programing for new request. 466 + * the programming for new request. 467 467 */ 468 468 if (status & TEGRA_APBDMA_STATUS_ISE_EOC) { 469 469 dev_err(tdc2dev(tdc),
+1 -1
drivers/dma/xgene-dma.c
··· 1742 1742 /* Initialize DMA channels software state */ 1743 1743 xgene_dma_init_channels(pdma); 1744 1744 1745 - /* Configue DMA rings */ 1745 + /* Configure DMA rings */ 1746 1746 ret = xgene_dma_init_rings(pdma); 1747 1747 if (ret) 1748 1748 goto err_clk_enable;
+2 -2
drivers/dma/xilinx/xilinx_dpdma.c
··· 149 149 * @addr_ext: upper 16 bit of 48 bit address (next_desc and src_addr) 150 150 * @next_desc: next descriptor 32 bit address 151 151 * @src_addr: payload source address (1st page, 32 LSB) 152 - * @addr_ext_23: payload source address (3nd and 3rd pages, 16 LSBs) 152 + * @addr_ext_23: payload source address (2nd and 3rd pages, 16 LSBs) 153 153 * @addr_ext_45: payload source address (4th and 5th pages, 16 LSBs) 154 154 * @src_addr2: payload source address (2nd page, 32 LSB) 155 155 * @src_addr3: payload source address (3rd page, 32 LSB) ··· 210 210 * @vchan: virtual DMA channel 211 211 * @reg: register base address 212 212 * @id: channel ID 213 - * @wait_to_stop: queue to wait for outstanding transacitons before stopping 213 + * @wait_to_stop: queue to wait for outstanding transactions before stopping 214 214 * @running: true if the channel is running 215 215 * @first_frame: flag for the first frame of stream 216 216 * @video_group: flag if multi-channel operation is needed for video channels