Re: [PATCH v2 01/29] nvmem: add support for cell lookups
From: Bartosz Golaszewski
Date: Tue Aug 28 2018 - 07:56:32 EST
2018-08-28 12:15 GMT+02:00 Srinivas Kandagatla <srinivas.kandagatla@xxxxxxxxxx>:
> On 27/08/18 14:37, Bartosz Golaszewski wrote:
>> I didn't notice it before but there's a global list of nvmem cells
> Bit of history here.
> The global list of nvmem_cell is to assist non device tree based cell
> lookups. These cell entries come as part of the non-dt providers
> All the device tree based cell lookup happen dynamically on request/demand,
> and all the cell definition comes from DT.
Makes perfect sense.
> As of today NVMEM supports both DT and non DT usecase, this is much simpler.
> Non dt cases have various consumer usecases.
> 1> Consumer is aware of provider name and cell details.
> This is probably simple usecase where it can just use device based
> 2> Consumer is not aware of provider name, its just aware of cell name.
> This is the case where global list of cells are used.
I would like to support an additional use case here: the provider is
generic and is not aware of its cells at all. Since the only way of
defining nvmem cells is through DT or nvmem_config, we lack a way to
allow machine code to define cells without the provider code being
>> with each cell referencing its owner nvmem device. I'm wondering if
>> this isn't some kind of inversion of ownership. Shouldn't each nvmem
>> device have a separate list of nvmem cells owned by it? What happens
> This is mainly done for use case where consumer does not have idea of
> provider name or any details.
It doesn't need to know the provider details, but in most subsystems
the core code associates such resources by dev_id and optional con_id
as Boris already said.
> First thing non dt user should do is use "NVMEM device based consumer APIs"
> ex: First get handle to nvmem device using its nvmem provider name by
> calling nvmem_device_get(); and use nvmem_device_cell_read/write() apis.
> Also am not 100% sure how would maintaining cells list per nvmem provider
> would help for the intended purpose of global list?
It would fix the use case where the consumer wants to use
nvmem_cell_get(dev, name) and two nvmem providers would have a cell
with the same name.
Next we could add a way to associate dev_ids with nvmem cells.
>> if we have two nvmem providers with the same names for cells? I'm
> Yes, it would return the first instance.. which is a known issue.
> Am not really sure this is a big problem as of today! but am open for any
> better suggestions!
Yes, I would like to rework nvmem a bit. I don't see any non-DT users
defining nvmem-cells using nvmem_config. I think that what we need is
a way of specifying cell config outside of nvmem providers in some
kind of structures. These tables would reference the provider by name
and define the cells. Then we would have an additional lookup
structure which would associate the consumer (by dev_id and con_id,
where dev_id could optionally be NULL and where we would fall back to
using con_id only) and the nvmem provider + cell together. Similarly
to how GPIO consumers are associated with the gpiochip and hwnum. How
does it sound?
>> asking because dev_id based lookup doesn't make sense if internally
>> nvmem_cell_get_from_list() doesn't care about any device names (takes
>> only the cell_id as argument).
> As I said this is for non DT usecase where consumers are not aware of
> provider details.
>> This doesn't cause any trouble now since there are no users defining
>> cells in nvmem_config - there are only DT users - but this must be
>> clarified before I can advance with correctly implementing nvmem
> DT users should not be defining this to start with! It's redundant and does
> not make sense!
Yes, this is what I said: we only seem to have DT users, so this API
is not used at the moment.
>> BTW: of_nvmem_cell_get() seems to always allocate an nvmem_cell
>> instance even if the cell for this node was already added to the nvmem
> I hope you got the reason why of_nvmem_cell_get() always allocates new
> instance for every get!!
I admit I didn't test it, but just from reading the code it seems like
in nvmem_cell_get() for DT-users we'll always get to
of_nvmem_cell_get() and in there we always end up calling line 873:
cell = kzalloc(sizeof(*cell), GFP_KERNEL);
There may be something I'm missing though.