@@ -146,7 +146,7 @@ struct tegra_dma_channel_regs {
146
146
};
147
147
148
148
/*
149
- * tegra_dma_sg_req: Dma request details to configure hardware. This
149
+ * tegra_dma_sg_req: DMA request details to configure hardware. This
150
150
* contains the details for one transfer to configure DMA hw.
151
151
* The client's request for data transfer can be broken into multiple
152
152
* sub-transfer as per requester details and hw support.
@@ -574,7 +574,7 @@ static bool handle_continuous_head_request(struct tegra_dma_channel *tdc,
574
574
struct tegra_dma_sg_req * hsgreq = NULL ;
575
575
576
576
if (list_empty (& tdc -> pending_sg_req )) {
577
- dev_err (tdc2dev (tdc ), "Dma is running without req\n" );
577
+ dev_err (tdc2dev (tdc ), "DMA is running without req\n" );
578
578
tegra_dma_stop (tdc );
579
579
return false;
580
580
}
@@ -587,7 +587,7 @@ static bool handle_continuous_head_request(struct tegra_dma_channel *tdc,
587
587
hsgreq = list_first_entry (& tdc -> pending_sg_req , typeof (* hsgreq ), node );
588
588
if (!hsgreq -> configured ) {
589
589
tegra_dma_stop (tdc );
590
- dev_err (tdc2dev (tdc ), "Error in dma transfer, aborting dma \n" );
590
+ dev_err (tdc2dev (tdc ), "Error in DMA transfer, aborting DMA \n" );
591
591
tegra_dma_abort_all (tdc );
592
592
return false;
593
593
}
@@ -922,7 +922,7 @@ static int get_transfer_param(struct tegra_dma_channel *tdc,
922
922
return 0 ;
923
923
924
924
default :
925
- dev_err (tdc2dev (tdc ), "Dma direction is not supported\n" );
925
+ dev_err (tdc2dev (tdc ), "DMA direction is not supported\n" );
926
926
return - EINVAL ;
927
927
}
928
928
return - EINVAL ;
@@ -955,7 +955,7 @@ static struct dma_async_tx_descriptor *tegra_dma_prep_slave_sg(
955
955
enum dma_slave_buswidth slave_bw ;
956
956
957
957
if (!tdc -> config_init ) {
958
- dev_err (tdc2dev (tdc ), "dma channel is not configured\n" );
958
+ dev_err (tdc2dev (tdc ), "DMA channel is not configured\n" );
959
959
return NULL ;
960
960
}
961
961
if (sg_len < 1 ) {
@@ -988,7 +988,7 @@ static struct dma_async_tx_descriptor *tegra_dma_prep_slave_sg(
988
988
989
989
dma_desc = tegra_dma_desc_get (tdc );
990
990
if (!dma_desc ) {
991
- dev_err (tdc2dev (tdc ), "Dma descriptors not available\n" );
991
+ dev_err (tdc2dev (tdc ), "DMA descriptors not available\n" );
992
992
return NULL ;
993
993
}
994
994
INIT_LIST_HEAD (& dma_desc -> tx_list );
@@ -1008,14 +1008,14 @@ static struct dma_async_tx_descriptor *tegra_dma_prep_slave_sg(
1008
1008
if ((len & 3 ) || (mem & 3 ) ||
1009
1009
(len > tdc -> tdma -> chip_data -> max_dma_count )) {
1010
1010
dev_err (tdc2dev (tdc ),
1011
- "Dma length/memory address is not supported\n" );
1011
+ "DMA length/memory address is not supported\n" );
1012
1012
tegra_dma_desc_put (tdc , dma_desc );
1013
1013
return NULL ;
1014
1014
}
1015
1015
1016
1016
sg_req = tegra_dma_sg_req_get (tdc );
1017
1017
if (!sg_req ) {
1018
- dev_err (tdc2dev (tdc ), "Dma sg-req not available\n" );
1018
+ dev_err (tdc2dev (tdc ), "DMA sg-req not available\n" );
1019
1019
tegra_dma_desc_put (tdc , dma_desc );
1020
1020
return NULL ;
1021
1021
}
@@ -1090,7 +1090,7 @@ static struct dma_async_tx_descriptor *tegra_dma_prep_dma_cyclic(
1090
1090
* terminating the DMA.
1091
1091
*/
1092
1092
if (tdc -> busy ) {
1093
- dev_err (tdc2dev (tdc ), "Request not allowed when dma running\n" );
1093
+ dev_err (tdc2dev (tdc ), "Request not allowed when DMA running\n" );
1094
1094
return NULL ;
1095
1095
}
1096
1096
@@ -1147,7 +1147,7 @@ static struct dma_async_tx_descriptor *tegra_dma_prep_dma_cyclic(
1147
1147
while (remain_len ) {
1148
1148
sg_req = tegra_dma_sg_req_get (tdc );
1149
1149
if (!sg_req ) {
1150
- dev_err (tdc2dev (tdc ), "Dma sg-req not available\n" );
1150
+ dev_err (tdc2dev (tdc ), "DMA sg-req not available\n" );
1151
1151
tegra_dma_desc_put (tdc , dma_desc );
1152
1152
return NULL ;
1153
1153
}
0 commit comments