RFC: Zynq Clock Controller
Sören Brinkmann
soren.brinkmann at xilinx.com
Fri Mar 8 12:38:59 EST 2013
On Fri, Mar 08, 2013 at 08:12:20AM +0100, Lars-Peter Clausen wrote:
> On 03/08/2013 12:25 AM, Sören Brinkmann wrote:
> > On Thu, Mar 07, 2013 at 11:02:58PM +0100, Lars-Peter Clausen wrote:
> >> On 03/07/2013 08:11 PM, Sören Brinkmann wrote:
> >>> On Thu, Mar 07, 2013 at 10:36:35AM +0100, Lars-Peter Clausen wrote:
> >>>> On 03/06/2013 06:27 PM, Sören Brinkmann wrote:
> >>>>> Hi Jan,
> >>>>>
> >>>>> what a small world. Good to hear from you.
> >>>>>
> >>>>> On Wed, Mar 06, 2013 at 12:51:21PM +0100, Jan Lübbe wrote:
> >>>>>> Hi Sören,
> >>>>>>
> >>>>>> On Tue, 2013-03-05 at 12:04 -0800, Sören Brinkmann wrote:
> >>>>>>> For this reasons, I'd like to propose moving Zynq into the same
> >>>>>>> direction. I.e. adding a clock controller with the following DT
> >>>>>>> description (details may change but the general idea should become
> >>>>>>> clear):
> >>>>>>> clkc: clkc {
> >>>>>>> #clock-cells = <1>;
> >>>>>>> compatible = "xlnx,ps7-clkc";
> >>>>>>> ps_clk_frequency = <33333333>; # board x-tal
> >>>>>>> # optional props
> >>>>>>> gem0_emio_clk_freq = <125000000>;
> >>>>>>> gem1_emio_clk_freq = <50000000>;
> >>>>>>> can_mio_clk_freq_xx = <1234>; # this is possible 54 times with xx = 00..53
> >>>>>>> };
> >>>>
> >>>> I definitely prefer the way it is right now in upstream, where we have one
> >>>> dt node per clock. It is more descriptive and also more extensible. And you
> >>>> also don't have to remember the clock index, and can use the phandle
> >>>> directly instead.
> >>> The problems I see with that are:
> >>> - with the clock controller we can model the clock tree as we need without changing the DT all the time
> >>
> >> When do we need to change the clock tree? The clock tree is pretty much fixed
> >> in hardware. And the DT describes the hardware, so I don't so a problem there.
> >>
> >>> - w/o a clock controller there is no block which has knowledge of the whole clock tree (unless we parse the DT), and could conduct reparent operations and similar
> >>
> >> The clk framework builds a representation of the clock tree. Each clock should
> >> be able to handle re-parenting on it's own, without knowing about the other
> >> clocks in the tree, the parent is selected by a field in the clocks register.
> >> It doesn't even have to know who the parents are, the clk framework will take
> >> care of all of this and just say, ok, switch your input clock to X, where X is
> >> a simple integer number The clk framework will also take care of re-calculating
> >> all the updates frequencies after re-parenting.
> > Nope, clk_set_parent() takes two 'struct *clk' as argument. So, you have to have those of the clock to change and all its parents. A device driver for example, which gets a clock through clk_get() does not have that information and should not, since it should not have to be aware of the SOC's clock hierarchy, IMHO.
> >
>
> Yes, you global clk_set_parent() method takes two clk structs. But the clock
> framework takes care of all the magic of mapping that second clk struct to a
> integer number, based on the clks parent list. So your set_parent callback in
> your clk_ops takes as parameters your clk and the index of the parent.
Right, but who/what in your scheme is supposed to call the high level API function requiring knowledge about all those 'struct *clk' and their hierarchy? This is where I think a big block has the advantage since it has a holistic view on the clock tree.
>
> >>
> >>> - once the clock controller is properly defined the clock connection should be contained in a dtsi which never changes. So, regarding remembering outputs, I don't see a problem. I rather see issues with having a pile of clocks described in the DT. I have currently 44 outputs proposed, and to model the whole tree a lot of more clocks are used (I started modeling everything with the clock primitives and removed custom clock implementations, except for the PLLs). Having all those in the DT does not really help, IMHO.
> >>
> >> Well, except if you want to use external clocks for ethernet or CAN. Also you
> >> don't need to change the dtsi either if you have a separate node for each clock.
> >>
> >> To avoid misunderstandings let me clarify that I don't want to have one node
> >> per clock, but rather one node per clock block (or whatever they are called).
> >> The combination of input select + divider + output gate. E.g. the UART clock
> >> block with it's 3 inputs and two outputs.
> >>
> >> Also the APER clocks should probably be one node with 24 outputs.
> > Okay, so instead of having one block encapsulating all clocks, you want it on a finer granularity. I don't see huge differences why that should be advantageous? It would just mean to create several blocks with their custom DT bindings instead of a single one. Just the abstraction level would be a little different.
>
> The DT is supposed to describe the hardware, not the software. These are the
> basic blocks. The are independent of each other and mostly orthogonal. It's the
> same IP block instantiated a couple of times next to each other (sometimes with
> slightly different parameters). They just happen to have their register mapped
> in the same IO region. The SLCR is just a container which contains all these
> blocks. Very much the same way the whole SoC is a container which contains the
> different peripheral blocks, etc. By you rationale we could argue, why do we
> need a dt node for each peripheral and not just simply one ZYNQ node?
I see us ending up here with custom clock blocks for each and every peripheral. They all look similar on first look but have all their slight differences. So, it's either different implementations or having a lot of DT properties to make a single implementation work for all the variations. That's why I prefer the single controller solution.
The DT description will always be an abstraction. You prefer grouping clocks for functional units, whereas I put them all into the same drawer labeled 'clocks'.
Sören
More information about the linux-arm-kernel
mailing list