[PATCH v2 1/2] staging: kpc2000: kpc_dma: rearrange lines exceeding 100 columns

Deepak R Varma mh12gx2825 at gmail.com
Mon Oct 26 04:04:53 UTC 2020


On Wed, Oct 21, 2020 at 01:01:07PM +0530, Deepak R Varma wrote:

Hello,
Requesting a review / ack of this patch.

Thank you,
Deepak.

> Reformat lines that exceed 100 column in length. Issue reported by
> checkpatch script.
> 
> Signed-off-by: Deepak R Varma <mh12gx2825 at gmail.com>
> ---
> Changes since v1:
>    - No change in this patch.
>    - Patch 2/2 has a change.
> 
>  drivers/staging/kpc2000/kpc_dma/dma.c         | 27 +++++++++---
>  drivers/staging/kpc2000/kpc_dma/fileops.c     | 44 +++++++++++++++----
>  .../staging/kpc2000/kpc_dma/kpc_dma_driver.c  |  9 ++--
>  3 files changed, 63 insertions(+), 17 deletions(-)
> 
> diff --git a/drivers/staging/kpc2000/kpc_dma/dma.c b/drivers/staging/kpc2000/kpc_dma/dma.c
> index 452a3f7c835d..b8d8294aa4c3 100644
> --- a/drivers/staging/kpc2000/kpc_dma/dma.c
> +++ b/drivers/staging/kpc2000/kpc_dma/dma.c
> @@ -16,7 +16,8 @@ irqreturn_t  ndd_irq_handler(int irq, void *dev_id)
>  {
>  	struct kpc_dma_device *ldev = (struct kpc_dma_device *)dev_id;
>  
> -	if ((GetEngineControl(ldev) & ENG_CTL_IRQ_ACTIVE) || (ldev->desc_completed->MyDMAAddr != GetEngineCompletePtr(ldev)))
> +	if ((GetEngineControl(ldev) & ENG_CTL_IRQ_ACTIVE) ||
> +	    (ldev->desc_completed->MyDMAAddr != GetEngineCompletePtr(ldev)))
>  		schedule_work(&ldev->irq_work);
>  
>  	return IRQ_HANDLED;
> @@ -39,7 +40,9 @@ void  ndd_irq_worker(struct work_struct *ws)
>  	cur = eng->desc_completed;
>  	do {
>  		cur = cur->Next;
> -		dev_dbg(&eng->pldev->dev, "Handling completed descriptor %p (acd = %p)\n", cur, cur->acd);
> +		dev_dbg(&eng->pldev->dev, "Handling completed descriptor %p (acd = %p)\n",
> +			cur,
> +			cur->acd);
>  		BUG_ON(cur == eng->desc_next); // Ordering failure.
>  
>  		if (cur->DescControlFlags & DMA_DESC_CTL_SOP) {
> @@ -56,7 +59,9 @@ void  ndd_irq_worker(struct work_struct *ws)
>  
>  		if (cur->DescControlFlags & DMA_DESC_CTL_EOP) {
>  			if (cur->acd)
> -				transfer_complete_cb(cur->acd, eng->accumulated_bytes, eng->accumulated_flags | ACD_FLAG_DONE);
> +				transfer_complete_cb(cur->acd,
> +						     eng->accumulated_bytes,
> +						     eng->accumulated_flags | ACD_FLAG_DONE);
>  		}
>  
>  		eng->desc_completed = cur;
> @@ -103,7 +108,10 @@ int  setup_dma_engine(struct kpc_dma_device *eng, u32 desc_cnt)
>  		eng->dir = DMA_TO_DEVICE;
>  
>  	eng->desc_pool_cnt = desc_cnt;
> -	eng->desc_pool = dma_pool_create("KPC DMA Descriptors", &eng->pldev->dev, sizeof(struct kpc_dma_descriptor), DMA_DESC_ALIGNMENT, 4096);
> +	eng->desc_pool = dma_pool_create("KPC DMA Descriptors",
> +					 &eng->pldev->dev,
> +					 sizeof(struct kpc_dma_descriptor),
> +					 DMA_DESC_ALIGNMENT, 4096);
>  
>  	eng->desc_pool_first = dma_pool_alloc(eng->desc_pool, GFP_KERNEL | GFP_DMA, &head_handle);
>  	if (!eng->desc_pool_first) {
> @@ -141,7 +149,11 @@ int  setup_dma_engine(struct kpc_dma_device *eng, u32 desc_cnt)
>  	INIT_WORK(&eng->irq_work, ndd_irq_worker);
>  
>  	// Grab IRQ line
> -	rv = request_irq(eng->irq, ndd_irq_handler, IRQF_SHARED, KP_DRIVER_NAME_DMA_CONTROLLER, eng);
> +	rv = request_irq(eng->irq,
> +			 ndd_irq_handler,
> +			 IRQF_SHARED,
> +			 KP_DRIVER_NAME_DMA_CONTROLLER,
> +			 eng);
>  	if (rv) {
>  		dev_err(&eng->pldev->dev, "%s: failed to request_irq: %d\n", __func__, rv);
>  		return rv;
> @@ -195,7 +207,10 @@ void  stop_dma_engine(struct kpc_dma_device *eng)
>  	}
>  
>  	// Clear any persistent bits just to make sure there is no residue from the reset
> -	SetClearEngineControl(eng, (ENG_CTL_IRQ_ACTIVE | ENG_CTL_DESC_COMPLETE | ENG_CTL_DESC_ALIGN_ERR | ENG_CTL_DESC_FETCH_ERR | ENG_CTL_SW_ABORT_ERR | ENG_CTL_DESC_CHAIN_END | ENG_CTL_DMA_WAITING_PERSIST), 0);
> +	SetClearEngineControl(eng, (ENG_CTL_IRQ_ACTIVE | ENG_CTL_DESC_COMPLETE |
> +				    ENG_CTL_DESC_ALIGN_ERR | ENG_CTL_DESC_FETCH_ERR |
> +				    ENG_CTL_SW_ABORT_ERR | ENG_CTL_DESC_CHAIN_END |
> +				    ENG_CTL_DMA_WAITING_PERSIST), 0);
>  
>  	// Reset performance counters
>  
> diff --git a/drivers/staging/kpc2000/kpc_dma/fileops.c b/drivers/staging/kpc2000/kpc_dma/fileops.c
> index e1c7c04f16fe..b929987844ff 100644
> --- a/drivers/staging/kpc2000/kpc_dma/fileops.c
> +++ b/drivers/staging/kpc2000/kpc_dma/fileops.c
> @@ -76,7 +76,11 @@ static int kpc_dma_transfer(struct dev_private_data *priv,
>  
>  	// Lock the user buffer pages in memory, and hold on to the page pointers (for the sglist)
>  	mmap_read_lock(current->mm);      /*  get memory map semaphore */
> -	rv = pin_user_pages(iov_base, acd->page_count, FOLL_TOUCH | FOLL_WRITE, acd->user_pages, NULL);
> +	rv = pin_user_pages(iov_base,
> +			    acd->page_count,
> +			    FOLL_TOUCH | FOLL_WRITE,
> +			    acd->user_pages,
> +			    NULL);
>  	mmap_read_unlock(current->mm);        /*  release the semaphore */
>  	if (rv != acd->page_count) {
>  		nr_pages = rv;
> @@ -89,16 +93,25 @@ static int kpc_dma_transfer(struct dev_private_data *priv,
>  	nr_pages = acd->page_count;
>  
>  	// Allocate and setup the sg_table (scatterlist entries)
> -	rv = sg_alloc_table_from_pages(&acd->sgt, acd->user_pages, acd->page_count, iov_base & (PAGE_SIZE - 1), iov_len, GFP_KERNEL);
> +	rv = sg_alloc_table_from_pages(&acd->sgt,
> +				       acd->user_pages,
> +				       acd->page_count,
> +				       iov_base & (PAGE_SIZE - 1),
> +				       iov_len, GFP_KERNEL);
>  	if (rv) {
>  		dev_err(&priv->ldev->pldev->dev, "Couldn't alloc sg_table (%d)\n", rv);
>  		goto unpin_pages;
>  	}
>  
>  	// Setup the DMA mapping for all the sg entries
> -	acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir);
> +	acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev,
> +					     acd->sgt.sgl,
> +					     acd->sgt.nents,
> +					     ldev->dir);
>  	if (acd->mapped_entry_count <= 0) {
> -		dev_err(&priv->ldev->pldev->dev, "Couldn't dma_map_sg (%d)\n", acd->mapped_entry_count);
> +		dev_err(&priv->ldev->pldev->dev,
> +			"Couldn't dma_map_sg (%d)\n",
> +			acd->mapped_entry_count);
>  		goto free_table;
>  	}
>  
> @@ -111,14 +124,26 @@ static int kpc_dma_transfer(struct dev_private_data *priv,
>  
>  	// Figoure out how many descriptors are available and return an error if there aren't enough
>  	num_descrs_avail = count_descriptors_available(ldev);
> -	dev_dbg(&priv->ldev->pldev->dev, "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d\n", acd->mapped_entry_count, desc_needed, num_descrs_avail);
> +	dev_dbg(&priv->ldev->pldev->dev,
> +		"    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d\n",
> +		acd->mapped_entry_count,
> +		desc_needed,
> +		num_descrs_avail);
>  	if (desc_needed >= ldev->desc_pool_cnt) {
> -		dev_warn(&priv->ldev->pldev->dev, "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    TOO MANY to ever complete!\n", acd->mapped_entry_count, desc_needed, num_descrs_avail);
> +		dev_warn(&priv->ldev->pldev->dev,
> +			 "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    TOO MANY to ever complete!\n",
> +			 acd->mapped_entry_count,
> +			 desc_needed,
> +			 num_descrs_avail);
>  		rv = -EAGAIN;
>  		goto err_descr_too_many;
>  	}
>  	if (desc_needed > num_descrs_avail) {
> -		dev_warn(&priv->ldev->pldev->dev, "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    Too many to complete right now.\n", acd->mapped_entry_count, desc_needed, num_descrs_avail);
> +		dev_warn(&priv->ldev->pldev->dev,
> +			 "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    Too many to complete right now.\n",
> +			 acd->mapped_entry_count,
> +			 desc_needed,
> +			 num_descrs_avail);
>  		rv = -EMSGSIZE;
>  		goto err_descr_too_many;
>  	}
> @@ -163,7 +188,10 @@ static int kpc_dma_transfer(struct dev_private_data *priv,
>  			if (i == acd->mapped_entry_count - 1 && p == pcnt - 1)
>  				desc->acd = acd;
>  
> -			dev_dbg(&priv->ldev->pldev->dev, "  Filled descriptor %p (acd = %p)\n", desc, desc->acd);
> +			dev_dbg(&priv->ldev->pldev->dev,
> +				"  Filled descriptor %p (acd = %p)\n",
> +				desc,
> +				desc->acd);
>  
>  			ldev->desc_next = desc->Next;
>  			desc = desc->Next;
> diff --git a/drivers/staging/kpc2000/kpc_dma/kpc_dma_driver.c b/drivers/staging/kpc2000/kpc_dma/kpc_dma_driver.c
> index 624d47bae4d1..7698e5ef2a7c 100644
> --- a/drivers/staging/kpc2000/kpc_dma/kpc_dma_driver.c
> +++ b/drivers/staging/kpc2000/kpc_dma/kpc_dma_driver.c
> @@ -138,7 +138,8 @@ int  kpc_dma_probe(struct platform_device *pldev)
>  
>  	// Setup miscdev struct
>  	dev = MKDEV(assigned_major_num, pldev->id);
> -	ldev->kpc_dma_dev = device_create(kpc_dma_class, &pldev->dev, dev, ldev, "kpc_dma%d", pldev->id);
> +	ldev->kpc_dma_dev = device_create(kpc_dma_class, &pldev->dev,
> +					  dev, ldev, "kpc_dma%d", pldev->id);
>  	if (IS_ERR(ldev->kpc_dma_dev)) {
>  		rv = PTR_ERR(ldev->kpc_dma_dev);
>  		dev_err(&ldev->pldev->dev, "%s: device_create failed: %d\n", __func__, rv);
> @@ -205,9 +206,11 @@ int __init kpc_dma_driver_init(void)
>  {
>  	int err;
>  
> -	err = __register_chrdev(KPC_DMA_CHAR_MAJOR, 0, KPC_DMA_NUM_MINORS, "kpc_dma", &kpc_dma_fops);
> +	err = __register_chrdev(KPC_DMA_CHAR_MAJOR, 0,
> +				KPC_DMA_NUM_MINORS, "kpc_dma", &kpc_dma_fops);
>  	if (err < 0) {
> -		pr_err("Can't allocate a major number (%d) for kpc_dma (err = %d)\n", KPC_DMA_CHAR_MAJOR, err);
> +		pr_err("Can't allocate a major number (%d) for kpc_dma (err = %d)\n",
> +		       KPC_DMA_CHAR_MAJOR, err);
>  		goto fail_chrdev_register;
>  	}
>  	assigned_major_num = err;
> -- 
> 2.25.1
> 


More information about the devel mailing list