@@ -420,9 +420,9 @@ ath11k_ce_alloc_ring(struct ath11k_base *sc, int nentries, int desc_sz)
* coherent DMA are unsupported
*/
ce_ring->base_addr_owner_space_unaligned =
- dma_zalloc_coherent(sc->dev,
- nentries * desc_sz + CE_DESC_RING_ALIGN,
- &base_addr, GFP_KERNEL);
+ dma_alloc_coherent(sc->dev,
+ nentries * desc_sz + CE_DESC_RING_ALIGN,
+ &base_addr, GFP_KERNEL);
if (!ce_ring->base_addr_owner_space_unaligned) {
kfree(ce_ring);
return ERR_PTR(-ENOMEM);
@@ -103,9 +103,9 @@ int ath11k_dp_srng_setup(struct ath11k_base *ab, struct dp_srng *ring,
num_entries = max_entries;
ring->size = (num_entries * entry_sz) + HAL_RING_BASE_ALIGN - 1;
- ring->vaddr_unaligned = dma_zalloc_coherent(ab->dev, ring->size,
- &ring->paddr_unaligned,
- GFP_KERNEL);
+ ring->vaddr_unaligned = dma_alloc_coherent(ab->dev, ring->size,
+ &ring->paddr_unaligned,
+ GFP_KERNEL);
if (!ring->vaddr_unaligned)
return -ENOMEM;
@@ -353,10 +353,9 @@ static int ath11k_dp_scatter_idle_link_desc_setup(struct ath11k_base *ab,
return -EINVAL;
for (i = 0; i < num_scatter_buf; i++) {
- slist[i].vaddr = dma_zalloc_coherent(
- ab->dev,
- HAL_WBM_IDLE_SCATTER_BUF_SIZE_MAX,
- &slist[i].paddr, GFP_KERNEL);
+ slist[i].vaddr = dma_alloc_coherent(ab->dev,
+ HAL_WBM_IDLE_SCATTER_BUF_SIZE_MAX,
+ &slist[i].paddr, GFP_KERNEL);
if (!slist[i].vaddr) {
ret = -ENOMEM;
goto err;
@@ -436,10 +435,9 @@ static int ath11k_dp_link_desc_bank_alloc(struct ath11k_base *ab,
if (i == (n_link_desc_bank - 1) && last_bank_sz)
desc_sz = last_bank_sz;
- desc_bank[i].vaddr_unaligned = dma_zalloc_coherent(
- ab->dev, desc_sz,
- &desc_bank[i].paddr_unaligned,
- GFP_KERNEL);
+ desc_bank[i].vaddr_unaligned = dma_alloc_coherent(ab->dev, desc_sz,
+ &desc_bank[i].paddr_unaligned,
+ GFP_KERNEL);
if (!desc_bank[i].vaddr_unaligned) {
ret = -ENOMEM;
goto err;