Re: [PATCH v9 07/22] clk: Add API to get index of the clock parent

From: Thierry Reding
Date: Fri Nov 08 2019 - 13:55:20 EST


On Fri, Nov 08, 2019 at 10:12:49AM -0800, Stephen Boyd wrote:
> Quoting Thierry Reding (2019-11-08 02:11:16)
> > On Thu, Nov 07, 2019 at 11:19:32AM -0800, Stephen Boyd wrote:
> > > Quoting Thierry Reding (2019-11-07 07:21:15)
> > > > On Thu, Nov 07, 2019 at 03:54:03AM +0300, Dmitry Osipenko wrote:
> > > > > 07.11.2019 02:10, Stephen Boyd ÐÐÑÐÑ:
> > > > > > Quoting Sowjanya Komatineni (2019-08-16 12:41:52)
> > > > > >> This patch adds an API clk_hw_get_parent_index to get index of the
> > > > > >> clock parent to use during the clock restore operations on system
> > > > > >> resume.
> > > > > >
> > > > > > Is there a reason we can't save the clk hw index at suspend time by
> > > > > > reading the hardware to understand the current parent? The parent index
> > > > > > typically doesn't matter unless we're trying to communicate something
> > > > > > from the framework to the provider driver. Put another way, I would
> > > > > > think the provider driver can figure out the index itself without having
> > > > > > to go through the framework to do so.
> > > > >
> > > > > Isn't it a bit wasteful to duplicate information about the parent within
> > > > > a provider if framework already has that info? The whole point of this
> > > > > new API is to allow providers to avoid that unnecessary duplication.
> > > > >
> > > > > Please note that clk_hw_get_parent_index is getting used only at the
> > > > > resume time and not at suspend.
> > > >
> > > > I agree with this. All of the information that we need is already cached
> > > > in the framework. Doing this in the driver would mean essentially adding
> > > > a "saved parent" field along with code to read the value at suspend time
> > > > to the three types of clocks that currently use this core helper.
> > >
> > > Don't we already have a "saved parent" field by storing the pointer to
> > > the clk_hw?
> > >
> > > >
> > > > That's certainly something that we *can* do, but it doesn't sound like a
> > > > better option than simply querying the framework for the value that we
> > > > need.
> > > >
> > >
> > > Let me say this another way. Why does this driver want to know the index
> > > that the framework uses for some clk_hw pointer? Perhaps it happens to
> > > align with the same value that hardware uses, but I still don't
> > > understand why the driver wants to know what the framework has decided
> > > is the index for some clk_hw pointer.
> > >
> > > Or is this something like "give me the index for the parent that the
> > > framework thinks I currently have but in reality don't have anymore
> > > because the register contents were wiped and we need to reparent it"?
> >
> > Yeah, that's exactly what this is being used for. It's used to restore
> > the parent/child relationship during resume after the registers have
> > been wiped during supend.
>
> Ok cool. Our whole suspend/resume and save/restore story hasn't really
> been well thought out so we may want to pull all this logic into the
> core one day. For now it's OK to do the heavy lifting from provider
> drivers until someone gets a better grasp on how this should all work.

Ah, that would explain why I was scratching my head trying to understand
how exactly this was supposed to work. It did feel like there was some
infrastructure there, but looking around there wasn't a very consistent
usage pattern that I could find.

I think suspend/resume is always a little tricky. For example the clocks
may required a slightly different logical sequences between SoCs. Maybe
even different types of clocks have different needs. We seem to have a
bit of that on Tegra alone already. Without having delved into this too
much, it seems to me like the core can't do a whole lot without stepping
(potentially) on drivers' toes.

The current save_context/restore_context seems to be mostly that,
though, so I think it's a good starting point. You're right that we may
eventually see clearer patterns appear.

> > > A generic API to get any index for this question is overkill and we should
> > > consider adding some sort of API like clk_hw_get_current_parent_index(),
> > > or a framework flag that tells the framework this parent is incorrect
> > > and we need to call the .set_parent() op again to reconfigure it.
> >
> > Okay, I think I see what you're saying. The current implementation does
> > carry a bit of a risk because users could be calling this function with
> > any arbitrary pair of struct clk_hw *, even completely unrelated ones.
> >
> > How about we turn it into this instead:
> >
> > /**
> > * clk_hw_get_parent_index - return the index of the parent clock
> > * @hw: clk_hw associated with the clk being consumed
> > *
> > * Fetches and returns the index of parent clock. Returns -EINVAL if the given
> > * clock does not have a current parent.
> > */
> > int clk_hw_get_parent_index(struct clk_hw *hw)
> > {
> > struct clk_hw *parent = clk_hw_get_parent(hw);
> >
> > if (!parent)
> > return -EINVAL;
> >
> > return clk_fetch_parent_index(hw->core, parent->core);
> > }
> > EXPORT_SYMBOL_GPL(clk_hw_get_parent_index);
> >
> > I think that has the advantage that we can't pass it a parent that's not
> > really a parent. There's still the slightly weird case where the clock
> > doesn't have a current parent, but hopefully that's something we are not
> > going to encounter much. After all this only makes sense to be called on
> > mux clocks and they always do have a parent by definition.
>
> Right.
>
> >
> > Perhaps we should be more explicit and wrap that !parent conditional in
> > a WARN_ON()? In my local patches I do that at the call sites because
> > they are all functions returning void, so we'd be silently ignoring the
> > cases, but I think it may make sense to have it in the core.
> >
>
> Sure a WARN_ON() sounds fair. That will not take the whole task down
> and makes sure that drivers aren't doing something incorrect. Otherwise,
> this looks good and we can optimize by caching the parent index later if
> we really need to.

Okay, great. I'll go replace the above patch in the branch that I have.
I'm not sure if you saw it, but I had sent this in a pull request for
v5.5-rc1 about a week ago because I've got Tegra clock driver patches
that depend on this. I can replace this patch with the above proposal
and update the Tegra clock driver branch and then resend the two pull
requests.

Does that sound like a plan?

Thierry

Attachment: signature.asc
Description: PGP signature