[RFC v5 06/17] dma-reserved-iommu: iommu_get/put_single_reserved
Eric Auger
eric.auger at linaro.org
Tue Mar 29 10:07:34 PDT 2016
Hi Jean-Philippe,
On 03/10/2016 12:52 PM, Jean-Philippe Brucker wrote:
> Hi Eric,
>
> On Tue, Mar 01, 2016 at 06:27:46PM +0000, Eric Auger wrote:
>> [...]
>> +
>> +int iommu_get_single_reserved(struct iommu_domain *domain,
>> + phys_addr_t addr, int prot,
>> + dma_addr_t *iova)
>> +{
>> + unsigned long order = __ffs(domain->ops->pgsize_bitmap);
>> + size_t page_size = 1 << order;
>> + phys_addr_t mask = page_size - 1;
>> + phys_addr_t aligned_addr = addr & ~mask;
>> + phys_addr_t offset = addr - aligned_addr;
>> + struct iommu_reserved_binding *b;
>> + struct iova *p_iova;
>> + struct iova_domain *iovad =
>> + (struct iova_domain *)domain->reserved_iova_cookie;
>> + int ret;
>> +
>> + if (!iovad)
>> + return -EINVAL;
>> +
>> + mutex_lock(&domain->reserved_mutex);
>
> I believe this function could get called from the chunk of __setup_irq
> that is executed atomically:
>
> * request_threaded_irq
> * __setup_irq
> * irq_startup
> * irq_domain_activate_irq
> * msi_domain_activate
> * msi_compose
> * iommu_get_single_reserved
>
> If this is the case, we should probably use a spinlock to protect the
> iova_domain...
Please apologize for the delay, I was in vacation.
Thank you for spotting this flow. I will rework the locking.
>
>> +
>> + b = find_reserved_binding(domain, aligned_addr, page_size);
>> + if (b) {
>> + *iova = b->iova + offset;
>> + kref_get(&b->kref);
>> + ret = 0;
>> + goto unlock;
>> + }
>> +
>> + /* there is no existing reserved iova for this pa */
>> + p_iova = alloc_iova(iovad, 1, iovad->dma_32bit_pfn, true);
>> + if (!p_iova) {
>> + ret = -ENOMEM;
>> + goto unlock;
>> + }
>> + *iova = p_iova->pfn_lo << order;
>> +
>> + b = kzalloc(sizeof(*b), GFP_KERNEL);
>
> ... and GFP_ATOMIC here.
OK
Thank you for your time!
Best Regards
Eric
>
> Thanks,
> Jean-Philippe
>
>> + if (!b) {
>> + ret = -ENOMEM;
>> + goto free_iova_unlock;
>> + }
>> +
>> + ret = iommu_map(domain, *iova, aligned_addr, page_size, prot);
>> + if (ret)
>> + goto free_binding_iova_unlock;
>> +
>> + kref_init(&b->kref);
>> + kref_get(&b->kref);
>> + b->domain = domain;
>> + b->addr = aligned_addr;
>> + b->iova = *iova;
>> + b->size = page_size;
>> +
>> + link_reserved_binding(domain, b);
>> +
>> + *iova += offset;
>> + goto unlock;
>> +
>> +free_binding_iova_unlock:
>> + kfree(b);
>> +free_iova_unlock:
>> + free_iova(iovad, *iova >> order);
>> +unlock:
>> + mutex_unlock(&domain->reserved_mutex);
>> + return ret;
>> +}
>> +EXPORT_SYMBOL_GPL(iommu_get_single_reserved);
More information about the linux-arm-kernel
mailing list