Re: [RFC 1/4] dma-buf: Add constraints sharing information
From: Daniel Vetter
Date: Wed Dec 10 2014 - 08:46:53 EST
On Wed, Dec 10, 2014 at 07:01:16PM +0530, Sumit Semwal wrote:
> Hi Daniel,
> Thanks a bunch for your review comments! A few comments, post our
> discussion at LPC;
> On 12 October 2014 at 00:25, Daniel Vetter <daniel@xxxxxxxx> wrote:
> > On Sat, Oct 11, 2014 at 01:37:55AM +0530, Sumit Semwal wrote:
> >> At present, struct device lacks a mechanism of exposing memory
> >> access constraints for the device.
> >> Consequently, there is also no mechanism to share these constraints
> >> while sharing buffers using dma-buf.
> >> If we add support for sharing such constraints, we could use that
> >> to try to collect requirements of different buffer-sharing devices
> >> to allocate buffers from a pool that satisfies requirements of all
> >> such devices.
> >> This is an attempt to add this support; at the moment, only a bitmask
> >> is added, but if post discussion, we realise we need more information,
> >> we could always extend the definition of constraint.
> >> A new dma-buf op is also added, to allow exporters to interpret or decide
> >> on constraint-masks on their own. A default implementation is provided to
> >> just AND (&) all the constraint-masks.
> >> What constitutes a constraint-mask could be left for interpretation on a
> >> per-platform basis, while defining some common masks.
> >> Signed-off-by: Sumit Semwal <sumit.semwal@xxxxxxxxxx>
> >> Cc: linux-kernel@xxxxxxxxxxxxxxx
> >> Cc: Greg Kroah-Hartman <gregkh@xxxxxxxxxxxxxxxxxxx>
> >> Cc: linux-media@xxxxxxxxxxxxxxx
> >> Cc: dri-devel@xxxxxxxxxxxxxxxxxxxxx
> >> Cc: linaro-mm-sig@xxxxxxxxxxxxxxxx
> > Just a few high-level comments, I'm between conference travel but
> > hopefully I can discuss this a bit at plumbers next week.
> > - I agree that for the insane specific cases we need something opaque like
> > the access constraints mask you propose here. But for the normal case I
> > think the existing dma constraints in dma_params would go a long way,
> > and I think we should look at Rob's RFC from aeons ago to solve those:
> > https://lkml.org/lkml/2012/7/19/285
> > With this we should be able to cover the allocation constraints of 90%
> > of all cases hopefully.
> > - I'm not sure whether an opaque bitmask is good enough really, I suspect
> > that we also need various priorities between different allocators. With
> > the option that some allocators are flat-out incompatible.
> Your/Rob's idea to figure out the constraints wrt max number of
> segments in the sg_list can provide, like you said, maybe 80-90% of
> the allocation constraints hopefully. The opaque mask should help for
> the remaining 'crazy' cases, so I'll be glad to merge Rob's and my
> approach on defining the constraints.
> I should think a little bit more about the priority idea that you
> propose here (and in another patch), but atm I am unable to see how
> that could help solve the finding-out-constraints problem.
> > - The big bummer imo with ION is that it fully side-steps, but this
> > proposal here also seems to add entirely new allocators. My rough idea
> This proposal does borrow this bit from ION, but once we have the
> required changes done in the dma api itself, the allocators can just
> become shims to the dma api allocators (eg dma_alloc_coherent etc) for
> cases where they can be used directly, while leaving provision for any
> crazy platform-specific allocators, without the userspace having to
> worry about it.
> > was that at allocate/attach time we iterate over all attached devices
> > like in Rob's patch and compute the most constrained allocation
> > requirements. Then we pick the underlying dma api allocator for these
> > constraints. That probably means that we need to open up the dma api a
> > bit. But I guess for a start we could simply try to allocate from the
> > most constrained device. Together with the opaque bits you propose here
> > we could even map additional crazy requirements like that an allocation
> > must come from a specific memory bank (provided by a special-purpose CMA
> > region). That might also mean that requirements are exclusive and no
> > allocation is possible.
> My idea was a little variation on what you said here - rather than do
> compute the most constraint allocation 'after' devices have attached
> (and right now, we don't really have a way to know that - but that's
> another point), I'd proposed to do the compute on each attach request,
> so the requesting drivers can know immediately if the attachment will
> not work for the other currently attached devices.
Well I said allocate/attach ;-) But yeah if we check at attach and reject
anything that doesn't work then there's no need to check again when
allocating, it /should/ work. But perhaps good to be paranoid and check
> > - I'm not sure we should allow drivers to override the access constraint
> > checks really - the dma_buf interfaces already provide this possibility
> > through the ->attach callback. In there exporters are allowed to reject
> > the attachment for any reason whatsover.
> This override the access constraint check is again meant only as a
> helper, but I could sure drop it.
> > - I think we should at least provide a helper implementation to allocate
> > dma-buffers for multiple devices using the dma constraints logic we
> > implement here. I think we should even go as far as providing a default
> > implementation for dma-bufs which uses dma_alloc_coherent and this new
> > dma contstraints computation code internally. This should be good enough
> Ok, my idea was to keep the allocation helpers separate from dma-buf
> framework - hence the cenalloc idea; if it seems like an extremely
> terrible approach to separate out helpers, I could try and do an RFC
> based on your idea.
Oh, I like helpers, it'd just put them into the dma-buf code and integrate
it directly instead of creating something separate.
> > for almost all devices, except those that do crazy stuff like swap
> > support of buffer objects (gem/ttm), virtual hardware buffers (vmwgfx)
> > or have other special needs (e.g. non-coherent buffers as speed
> > optimization).
> Cenalloc type of idea could allow for these special needs I think!
Well imo we should aim for 90% first, fix out fallout and then reasses
what's needed. Tends to leat to better design overall.
Software Engineer, Intel Corporation
+41 (0) 79 365 57 48 - http://blog.ffwll.ch
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/