Merge tag 'kbuild-v5.15' of git://git.kernel.org/pub/scm/linux/kernel/git/masahiroy...
[linux-2.6-microblaze.git] / drivers / net / phy / nxp-c45-tja11xx.c
1 // SPDX-License-Identifier: GPL-2.0
2 /* NXP C45 PHY driver
3  * Copyright (C) 2021 NXP
4  * Author: Radu Pirea <radu-nicolae.pirea@oss.nxp.com>
5  */
6
7 #include <linux/delay.h>
8 #include <linux/ethtool.h>
9 #include <linux/ethtool_netlink.h>
10 #include <linux/kernel.h>
11 #include <linux/mii.h>
12 #include <linux/module.h>
13 #include <linux/phy.h>
14 #include <linux/processor.h>
15 #include <linux/property.h>
16 #include <linux/ptp_classify.h>
17 #include <linux/ptp_clock_kernel.h>
18 #include <linux/net_tstamp.h>
19
20 #define PHY_ID_TJA_1103                 0x001BB010
21
22 #define PMAPMD_B100T1_PMAPMD_CTL        0x0834
23 #define B100T1_PMAPMD_CONFIG_EN         BIT(15)
24 #define B100T1_PMAPMD_MASTER            BIT(14)
25 #define MASTER_MODE                     (B100T1_PMAPMD_CONFIG_EN | \
26                                          B100T1_PMAPMD_MASTER)
27 #define SLAVE_MODE                      (B100T1_PMAPMD_CONFIG_EN)
28
29 #define VEND1_DEVICE_CONTROL            0x0040
30 #define DEVICE_CONTROL_RESET            BIT(15)
31 #define DEVICE_CONTROL_CONFIG_GLOBAL_EN BIT(14)
32 #define DEVICE_CONTROL_CONFIG_ALL_EN    BIT(13)
33
34 #define VEND1_PHY_IRQ_ACK               0x80A0
35 #define VEND1_PHY_IRQ_EN                0x80A1
36 #define VEND1_PHY_IRQ_STATUS            0x80A2
37 #define PHY_IRQ_LINK_EVENT              BIT(1)
38
39 #define VEND1_PHY_CONTROL               0x8100
40 #define PHY_CONFIG_EN                   BIT(14)
41 #define PHY_START_OP                    BIT(0)
42
43 #define VEND1_PHY_CONFIG                0x8108
44 #define PHY_CONFIG_AUTO                 BIT(0)
45
46 #define VEND1_SIGNAL_QUALITY            0x8320
47 #define SQI_VALID                       BIT(14)
48 #define SQI_MASK                        GENMASK(2, 0)
49 #define MAX_SQI                         SQI_MASK
50
51 #define VEND1_CABLE_TEST                0x8330
52 #define CABLE_TEST_ENABLE               BIT(15)
53 #define CABLE_TEST_START                BIT(14)
54 #define CABLE_TEST_VALID                BIT(13)
55 #define CABLE_TEST_OK                   0x00
56 #define CABLE_TEST_SHORTED              0x01
57 #define CABLE_TEST_OPEN                 0x02
58 #define CABLE_TEST_UNKNOWN              0x07
59
60 #define VEND1_PORT_CONTROL              0x8040
61 #define PORT_CONTROL_EN                 BIT(14)
62
63 #define VEND1_PORT_ABILITIES            0x8046
64 #define PTP_ABILITY                     BIT(3)
65
66 #define VEND1_PORT_INFRA_CONTROL        0xAC00
67 #define PORT_INFRA_CONTROL_EN           BIT(14)
68
69 #define VEND1_RXID                      0xAFCC
70 #define VEND1_TXID                      0xAFCD
71 #define ID_ENABLE                       BIT(15)
72
73 #define VEND1_ABILITIES                 0xAFC4
74 #define RGMII_ID_ABILITY                BIT(15)
75 #define RGMII_ABILITY                   BIT(14)
76 #define RMII_ABILITY                    BIT(10)
77 #define REVMII_ABILITY                  BIT(9)
78 #define MII_ABILITY                     BIT(8)
79 #define SGMII_ABILITY                   BIT(0)
80
81 #define VEND1_MII_BASIC_CONFIG          0xAFC6
82 #define MII_BASIC_CONFIG_REV            BIT(8)
83 #define MII_BASIC_CONFIG_SGMII          0x9
84 #define MII_BASIC_CONFIG_RGMII          0x7
85 #define MII_BASIC_CONFIG_RMII           0x5
86 #define MII_BASIC_CONFIG_MII            0x4
87
88 #define VEND1_SYMBOL_ERROR_COUNTER      0x8350
89 #define VEND1_LINK_DROP_COUNTER         0x8352
90 #define VEND1_LINK_LOSSES_AND_FAILURES  0x8353
91 #define VEND1_R_GOOD_FRAME_CNT          0xA950
92 #define VEND1_R_BAD_FRAME_CNT           0xA952
93 #define VEND1_R_RXER_FRAME_CNT          0xA954
94 #define VEND1_RX_PREAMBLE_COUNT         0xAFCE
95 #define VEND1_TX_PREAMBLE_COUNT         0xAFCF
96 #define VEND1_RX_IPG_LENGTH             0xAFD0
97 #define VEND1_TX_IPG_LENGTH             0xAFD1
98 #define COUNTER_EN                      BIT(15)
99
100 #define VEND1_LTC_LOAD_CTRL             0x1105
101 #define READ_LTC                        BIT(2)
102 #define LOAD_LTC                        BIT(0)
103
104 #define VEND1_LTC_WR_NSEC_0             0x1106
105 #define VEND1_LTC_WR_NSEC_1             0x1107
106 #define VEND1_LTC_WR_SEC_0              0x1108
107 #define VEND1_LTC_WR_SEC_1              0x1109
108
109 #define VEND1_LTC_RD_NSEC_0             0x110A
110 #define VEND1_LTC_RD_NSEC_1             0x110B
111 #define VEND1_LTC_RD_SEC_0              0x110C
112 #define VEND1_LTC_RD_SEC_1              0x110D
113
114 #define VEND1_RATE_ADJ_SUBNS_0          0x110F
115 #define VEND1_RATE_ADJ_SUBNS_1          0x1110
116 #define CLK_RATE_ADJ_LD                 BIT(15)
117 #define CLK_RATE_ADJ_DIR                BIT(14)
118
119 #define VEND1_HW_LTC_LOCK_CTRL          0x1115
120 #define HW_LTC_LOCK_EN                  BIT(0)
121
122 #define VEND1_PTP_IRQ_EN                0x1131
123 #define VEND1_PTP_IRQ_STATUS            0x1132
124 #define PTP_IRQ_EGR_TS                  BIT(0)
125
126 #define VEND1_RX_TS_INSRT_CTRL          0x114D
127 #define RX_TS_INSRT_MODE2               0x02
128
129 #define VEND1_EGR_RING_DATA_0           0x114E
130 #define VEND1_EGR_RING_DATA_1_SEQ_ID    0x114F
131 #define VEND1_EGR_RING_DATA_2_NSEC_15_0 0x1150
132 #define VEND1_EGR_RING_DATA_3           0x1151
133 #define VEND1_EGR_RING_CTRL             0x1154
134
135 #define RING_DATA_0_DOMAIN_NUMBER       GENMASK(7, 0)
136 #define RING_DATA_0_MSG_TYPE            GENMASK(11, 8)
137 #define RING_DATA_0_SEC_4_2             GENMASK(14, 2)
138 #define RING_DATA_0_TS_VALID            BIT(15)
139
140 #define RING_DATA_3_NSEC_29_16          GENMASK(13, 0)
141 #define RING_DATA_3_SEC_1_0             GENMASK(15, 14)
142 #define RING_DATA_5_SEC_16_5            GENMASK(15, 4)
143 #define RING_DONE                       BIT(0)
144
145 #define TS_SEC_MASK                     GENMASK(1, 0)
146
147 #define VEND1_PORT_FUNC_ENABLES         0x8048
148 #define PTP_ENABLE                      BIT(3)
149
150 #define VEND1_PORT_PTP_CONTROL          0x9000
151 #define PORT_PTP_CONTROL_BYPASS         BIT(11)
152
153 #define VEND1_PTP_CLK_PERIOD            0x1104
154 #define PTP_CLK_PERIOD_100BT1           15ULL
155
156 #define VEND1_EVENT_MSG_FILT            0x1148
157 #define EVENT_MSG_FILT_ALL              0x0F
158 #define EVENT_MSG_FILT_NONE             0x00
159
160 #define VEND1_TX_PIPE_DLY_NS            0x1149
161 #define VEND1_TX_PIPEDLY_SUBNS          0x114A
162 #define VEND1_RX_PIPE_DLY_NS            0x114B
163 #define VEND1_RX_PIPEDLY_SUBNS          0x114C
164
165 #define RGMII_PERIOD_PS                 8000U
166 #define PS_PER_DEGREE                   div_u64(RGMII_PERIOD_PS, 360)
167 #define MIN_ID_PS                       1644U
168 #define MAX_ID_PS                       2260U
169 #define DEFAULT_ID_PS                   2000U
170
171 #define PPM_TO_SUBNS_INC(ppb)   div_u64(GENMASK(31, 0) * (ppb) * \
172                                         PTP_CLK_PERIOD_100BT1, NSEC_PER_SEC)
173
174 #define NXP_C45_SKB_CB(skb)     ((struct nxp_c45_skb_cb *)(skb)->cb)
175
176 struct nxp_c45_skb_cb {
177         struct ptp_header *header;
178         unsigned int type;
179 };
180
181 struct nxp_c45_hwts {
182         u32     nsec;
183         u32     sec;
184         u8      domain_number;
185         u16     sequence_id;
186         u8      msg_type;
187 };
188
189 struct nxp_c45_phy {
190         struct phy_device *phydev;
191         struct mii_timestamper mii_ts;
192         struct ptp_clock *ptp_clock;
193         struct ptp_clock_info caps;
194         struct sk_buff_head tx_queue;
195         struct sk_buff_head rx_queue;
196         /* used to access the PTP registers atomic */
197         struct mutex ptp_lock;
198         int hwts_tx;
199         int hwts_rx;
200         u32 tx_delay;
201         u32 rx_delay;
202 };
203
204 struct nxp_c45_phy_stats {
205         const char      *name;
206         u8              mmd;
207         u16             reg;
208         u8              off;
209         u16             mask;
210 };
211
212 static bool nxp_c45_poll_txts(struct phy_device *phydev)
213 {
214         return phydev->irq <= 0;
215 }
216
217 static int _nxp_c45_ptp_gettimex64(struct ptp_clock_info *ptp,
218                                    struct timespec64 *ts,
219                                    struct ptp_system_timestamp *sts)
220 {
221         struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
222
223         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_LOAD_CTRL,
224                       READ_LTC);
225         ts->tv_nsec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
226                                    VEND1_LTC_RD_NSEC_0);
227         ts->tv_nsec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
228                                     VEND1_LTC_RD_NSEC_1) << 16;
229         ts->tv_sec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
230                                   VEND1_LTC_RD_SEC_0);
231         ts->tv_sec |= phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
232                                    VEND1_LTC_RD_SEC_1) << 16;
233
234         return 0;
235 }
236
237 static int nxp_c45_ptp_gettimex64(struct ptp_clock_info *ptp,
238                                   struct timespec64 *ts,
239                                   struct ptp_system_timestamp *sts)
240 {
241         struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
242
243         mutex_lock(&priv->ptp_lock);
244         _nxp_c45_ptp_gettimex64(ptp, ts, sts);
245         mutex_unlock(&priv->ptp_lock);
246
247         return 0;
248 }
249
250 static int _nxp_c45_ptp_settime64(struct ptp_clock_info *ptp,
251                                   const struct timespec64 *ts)
252 {
253         struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
254
255         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_NSEC_0,
256                       ts->tv_nsec);
257         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_NSEC_1,
258                       ts->tv_nsec >> 16);
259         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_SEC_0,
260                       ts->tv_sec);
261         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_WR_SEC_1,
262                       ts->tv_sec >> 16);
263         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_LTC_LOAD_CTRL,
264                       LOAD_LTC);
265
266         return 0;
267 }
268
269 static int nxp_c45_ptp_settime64(struct ptp_clock_info *ptp,
270                                  const struct timespec64 *ts)
271 {
272         struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
273
274         mutex_lock(&priv->ptp_lock);
275         _nxp_c45_ptp_settime64(ptp, ts);
276         mutex_unlock(&priv->ptp_lock);
277
278         return 0;
279 }
280
281 static int nxp_c45_ptp_adjfine(struct ptp_clock_info *ptp, long scaled_ppm)
282 {
283         struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
284         s32 ppb = scaled_ppm_to_ppb(scaled_ppm);
285         u64 subns_inc_val;
286         bool inc;
287
288         mutex_lock(&priv->ptp_lock);
289         inc = ppb >= 0;
290         ppb = abs(ppb);
291
292         subns_inc_val = PPM_TO_SUBNS_INC(ppb);
293
294         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_RATE_ADJ_SUBNS_0,
295                       subns_inc_val);
296         subns_inc_val >>= 16;
297         subns_inc_val |= CLK_RATE_ADJ_LD;
298         if (inc)
299                 subns_inc_val |= CLK_RATE_ADJ_DIR;
300
301         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_RATE_ADJ_SUBNS_1,
302                       subns_inc_val);
303         mutex_unlock(&priv->ptp_lock);
304
305         return 0;
306 }
307
308 static int nxp_c45_ptp_adjtime(struct ptp_clock_info *ptp, s64 delta)
309 {
310         struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
311         struct timespec64 now, then;
312
313         mutex_lock(&priv->ptp_lock);
314         then = ns_to_timespec64(delta);
315         _nxp_c45_ptp_gettimex64(ptp, &now, NULL);
316         now = timespec64_add(now, then);
317         _nxp_c45_ptp_settime64(ptp, &now);
318         mutex_unlock(&priv->ptp_lock);
319
320         return 0;
321 }
322
323 static void nxp_c45_reconstruct_ts(struct timespec64 *ts,
324                                    struct nxp_c45_hwts *hwts)
325 {
326         ts->tv_nsec = hwts->nsec;
327         if ((ts->tv_sec & TS_SEC_MASK) < (hwts->sec & TS_SEC_MASK))
328                 ts->tv_sec -= TS_SEC_MASK + 1;
329         ts->tv_sec &= ~TS_SEC_MASK;
330         ts->tv_sec |= hwts->sec & TS_SEC_MASK;
331 }
332
333 static bool nxp_c45_match_ts(struct ptp_header *header,
334                              struct nxp_c45_hwts *hwts,
335                              unsigned int type)
336 {
337         return ntohs(header->sequence_id) == hwts->sequence_id &&
338                ptp_get_msgtype(header, type) == hwts->msg_type &&
339                header->domain_number  == hwts->domain_number;
340 }
341
342 static bool nxp_c45_get_hwtxts(struct nxp_c45_phy *priv,
343                                struct nxp_c45_hwts *hwts)
344 {
345         bool valid;
346         u16 reg;
347
348         mutex_lock(&priv->ptp_lock);
349         phy_write_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_CTRL,
350                       RING_DONE);
351         reg = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_DATA_0);
352         valid = !!(reg & RING_DATA_0_TS_VALID);
353         if (!valid)
354                 goto nxp_c45_get_hwtxts_out;
355
356         hwts->domain_number = reg;
357         hwts->msg_type = (reg & RING_DATA_0_MSG_TYPE) >> 8;
358         hwts->sec = (reg & RING_DATA_0_SEC_4_2) >> 10;
359         hwts->sequence_id = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
360                                          VEND1_EGR_RING_DATA_1_SEQ_ID);
361         hwts->nsec = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1,
362                                   VEND1_EGR_RING_DATA_2_NSEC_15_0);
363         reg = phy_read_mmd(priv->phydev, MDIO_MMD_VEND1, VEND1_EGR_RING_DATA_3);
364         hwts->nsec |= (reg & RING_DATA_3_NSEC_29_16) << 16;
365         hwts->sec |= (reg & RING_DATA_3_SEC_1_0) >> 14;
366
367 nxp_c45_get_hwtxts_out:
368         mutex_unlock(&priv->ptp_lock);
369         return valid;
370 }
371
372 static void nxp_c45_process_txts(struct nxp_c45_phy *priv,
373                                  struct nxp_c45_hwts *txts)
374 {
375         struct sk_buff *skb, *tmp, *skb_match = NULL;
376         struct skb_shared_hwtstamps shhwtstamps;
377         struct timespec64 ts;
378         unsigned long flags;
379         bool ts_match;
380         s64 ts_ns;
381
382         spin_lock_irqsave(&priv->tx_queue.lock, flags);
383         skb_queue_walk_safe(&priv->tx_queue, skb, tmp) {
384                 ts_match = nxp_c45_match_ts(NXP_C45_SKB_CB(skb)->header, txts,
385                                             NXP_C45_SKB_CB(skb)->type);
386                 if (!ts_match)
387                         continue;
388                 skb_match = skb;
389                 __skb_unlink(skb, &priv->tx_queue);
390                 break;
391         }
392         spin_unlock_irqrestore(&priv->tx_queue.lock, flags);
393
394         if (skb_match) {
395                 nxp_c45_ptp_gettimex64(&priv->caps, &ts, NULL);
396                 nxp_c45_reconstruct_ts(&ts, txts);
397                 memset(&shhwtstamps, 0, sizeof(shhwtstamps));
398                 ts_ns = timespec64_to_ns(&ts);
399                 shhwtstamps.hwtstamp = ns_to_ktime(ts_ns);
400                 skb_complete_tx_timestamp(skb_match, &shhwtstamps);
401         } else {
402                 phydev_warn(priv->phydev,
403                             "the tx timestamp doesn't match with any skb\n");
404         }
405 }
406
407 static long nxp_c45_do_aux_work(struct ptp_clock_info *ptp)
408 {
409         struct nxp_c45_phy *priv = container_of(ptp, struct nxp_c45_phy, caps);
410         bool poll_txts = nxp_c45_poll_txts(priv->phydev);
411         struct skb_shared_hwtstamps *shhwtstamps_rx;
412         struct nxp_c45_hwts hwts;
413         bool reschedule = false;
414         struct timespec64 ts;
415         struct sk_buff *skb;
416         bool txts_valid;
417         u32 ts_raw;
418
419         while (!skb_queue_empty_lockless(&priv->tx_queue) && poll_txts) {
420                 txts_valid = nxp_c45_get_hwtxts(priv, &hwts);
421                 if (unlikely(!txts_valid)) {
422                         /* Still more skbs in the queue */
423                         reschedule = true;
424                         break;
425                 }
426
427                 nxp_c45_process_txts(priv, &hwts);
428         }
429
430         while ((skb = skb_dequeue(&priv->rx_queue)) != NULL) {
431                 nxp_c45_ptp_gettimex64(&priv->caps, &ts, NULL);
432                 ts_raw = __be32_to_cpu(NXP_C45_SKB_CB(skb)->header->reserved2);
433                 hwts.sec = ts_raw >> 30;
434                 hwts.nsec = ts_raw & GENMASK(29, 0);
435                 nxp_c45_reconstruct_ts(&ts, &hwts);
436                 shhwtstamps_rx = skb_hwtstamps(skb);
437                 shhwtstamps_rx->hwtstamp = ns_to_ktime(timespec64_to_ns(&ts));
438                 NXP_C45_SKB_CB(skb)->header->reserved2 = 0;
439                 netif_rx_ni(skb);
440         }
441
442         return reschedule ? 1 : -1;
443 }
444
445 static int nxp_c45_init_ptp_clock(struct nxp_c45_phy *priv)
446 {
447         priv->caps = (struct ptp_clock_info) {
448                 .owner          = THIS_MODULE,
449                 .name           = "NXP C45 PHC",
450                 .max_adj        = 16666666,
451                 .adjfine        = nxp_c45_ptp_adjfine,
452                 .adjtime        = nxp_c45_ptp_adjtime,
453                 .gettimex64     = nxp_c45_ptp_gettimex64,
454                 .settime64      = nxp_c45_ptp_settime64,
455                 .do_aux_work    = nxp_c45_do_aux_work,
456         };
457
458         priv->ptp_clock = ptp_clock_register(&priv->caps,
459                                              &priv->phydev->mdio.dev);
460
461         if (IS_ERR(priv->ptp_clock))
462                 return PTR_ERR(priv->ptp_clock);
463
464         if (!priv->ptp_clock)
465                 return -ENOMEM;
466
467         return 0;
468 }
469
470 static void nxp_c45_txtstamp(struct mii_timestamper *mii_ts,
471                              struct sk_buff *skb, int type)
472 {
473         struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
474                                                 mii_ts);
475
476         switch (priv->hwts_tx) {
477         case HWTSTAMP_TX_ON:
478                 NXP_C45_SKB_CB(skb)->type = type;
479                 NXP_C45_SKB_CB(skb)->header = ptp_parse_header(skb, type);
480                 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS;
481                 skb_queue_tail(&priv->tx_queue, skb);
482                 if (nxp_c45_poll_txts(priv->phydev))
483                         ptp_schedule_worker(priv->ptp_clock, 0);
484                 break;
485         case HWTSTAMP_TX_OFF:
486         default:
487                 kfree_skb(skb);
488                 break;
489         }
490 }
491
492 static bool nxp_c45_rxtstamp(struct mii_timestamper *mii_ts,
493                              struct sk_buff *skb, int type)
494 {
495         struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
496                                                 mii_ts);
497         struct ptp_header *header = ptp_parse_header(skb, type);
498
499         if (!header)
500                 return false;
501
502         if (!priv->hwts_rx)
503                 return false;
504
505         NXP_C45_SKB_CB(skb)->header = header;
506         skb_queue_tail(&priv->rx_queue, skb);
507         ptp_schedule_worker(priv->ptp_clock, 0);
508
509         return true;
510 }
511
512 static int nxp_c45_hwtstamp(struct mii_timestamper *mii_ts,
513                             struct ifreq *ifreq)
514 {
515         struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
516                                                 mii_ts);
517         struct phy_device *phydev = priv->phydev;
518         struct hwtstamp_config cfg;
519
520         if (copy_from_user(&cfg, ifreq->ifr_data, sizeof(cfg)))
521                 return -EFAULT;
522
523         if (cfg.tx_type < 0 || cfg.tx_type > HWTSTAMP_TX_ON)
524                 return -ERANGE;
525
526         priv->hwts_tx = cfg.tx_type;
527
528         switch (cfg.rx_filter) {
529         case HWTSTAMP_FILTER_NONE:
530                 priv->hwts_rx = 0;
531                 break;
532         case HWTSTAMP_FILTER_PTP_V2_L2_EVENT:
533         case HWTSTAMP_FILTER_PTP_V2_L2_SYNC:
534         case HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ:
535                 priv->hwts_rx = 1;
536                 cfg.rx_filter = HWTSTAMP_FILTER_PTP_V2_L2_EVENT;
537                 break;
538         default:
539                 return -ERANGE;
540         }
541
542         if (priv->hwts_rx || priv->hwts_tx) {
543                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_EVENT_MSG_FILT,
544                               EVENT_MSG_FILT_ALL);
545                 phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
546                                    VEND1_PORT_PTP_CONTROL,
547                                    PORT_PTP_CONTROL_BYPASS);
548         } else {
549                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_EVENT_MSG_FILT,
550                               EVENT_MSG_FILT_NONE);
551                 phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_PTP_CONTROL,
552                                  PORT_PTP_CONTROL_BYPASS);
553         }
554
555         if (nxp_c45_poll_txts(priv->phydev))
556                 goto nxp_c45_no_ptp_irq;
557
558         if (priv->hwts_tx)
559                 phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
560                                  VEND1_PTP_IRQ_EN, PTP_IRQ_EGR_TS);
561         else
562                 phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
563                                    VEND1_PTP_IRQ_EN, PTP_IRQ_EGR_TS);
564
565 nxp_c45_no_ptp_irq:
566         return copy_to_user(ifreq->ifr_data, &cfg, sizeof(cfg)) ? -EFAULT : 0;
567 }
568
569 static int nxp_c45_ts_info(struct mii_timestamper *mii_ts,
570                            struct ethtool_ts_info *ts_info)
571 {
572         struct nxp_c45_phy *priv = container_of(mii_ts, struct nxp_c45_phy,
573                                                 mii_ts);
574
575         ts_info->so_timestamping = SOF_TIMESTAMPING_TX_HARDWARE |
576                         SOF_TIMESTAMPING_RX_HARDWARE |
577                         SOF_TIMESTAMPING_RAW_HARDWARE;
578         ts_info->phc_index = ptp_clock_index(priv->ptp_clock);
579         ts_info->tx_types = (1 << HWTSTAMP_TX_OFF) | (1 << HWTSTAMP_TX_ON);
580         ts_info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) |
581                         (1 << HWTSTAMP_FILTER_PTP_V2_L2_SYNC) |
582                         (1 << HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ) |
583                         (1 << HWTSTAMP_FILTER_PTP_V2_L2_EVENT);
584
585         return 0;
586 }
587
588 static const struct nxp_c45_phy_stats nxp_c45_hw_stats[] = {
589         { "phy_symbol_error_cnt", MDIO_MMD_VEND1,
590                 VEND1_SYMBOL_ERROR_COUNTER, 0, GENMASK(15, 0) },
591         { "phy_link_status_drop_cnt", MDIO_MMD_VEND1,
592                 VEND1_LINK_DROP_COUNTER, 8, GENMASK(13, 8) },
593         { "phy_link_availability_drop_cnt", MDIO_MMD_VEND1,
594                 VEND1_LINK_DROP_COUNTER, 0, GENMASK(5, 0) },
595         { "phy_link_loss_cnt", MDIO_MMD_VEND1,
596                 VEND1_LINK_LOSSES_AND_FAILURES, 10, GENMASK(15, 10) },
597         { "phy_link_failure_cnt", MDIO_MMD_VEND1,
598                 VEND1_LINK_LOSSES_AND_FAILURES, 0, GENMASK(9, 0) },
599         { "r_good_frame_cnt", MDIO_MMD_VEND1,
600                 VEND1_R_GOOD_FRAME_CNT, 0, GENMASK(15, 0) },
601         { "r_bad_frame_cnt", MDIO_MMD_VEND1,
602                 VEND1_R_BAD_FRAME_CNT, 0, GENMASK(15, 0) },
603         { "r_rxer_frame_cnt", MDIO_MMD_VEND1,
604                 VEND1_R_RXER_FRAME_CNT, 0, GENMASK(15, 0) },
605         { "rx_preamble_count", MDIO_MMD_VEND1,
606                 VEND1_RX_PREAMBLE_COUNT, 0, GENMASK(5, 0) },
607         { "tx_preamble_count", MDIO_MMD_VEND1,
608                 VEND1_TX_PREAMBLE_COUNT, 0, GENMASK(5, 0) },
609         { "rx_ipg_length", MDIO_MMD_VEND1,
610                 VEND1_RX_IPG_LENGTH, 0, GENMASK(8, 0) },
611         { "tx_ipg_length", MDIO_MMD_VEND1,
612                 VEND1_TX_IPG_LENGTH, 0, GENMASK(8, 0) },
613 };
614
615 static int nxp_c45_get_sset_count(struct phy_device *phydev)
616 {
617         return ARRAY_SIZE(nxp_c45_hw_stats);
618 }
619
620 static void nxp_c45_get_strings(struct phy_device *phydev, u8 *data)
621 {
622         size_t i;
623
624         for (i = 0; i < ARRAY_SIZE(nxp_c45_hw_stats); i++) {
625                 strncpy(data + i * ETH_GSTRING_LEN,
626                         nxp_c45_hw_stats[i].name, ETH_GSTRING_LEN);
627         }
628 }
629
630 static void nxp_c45_get_stats(struct phy_device *phydev,
631                               struct ethtool_stats *stats, u64 *data)
632 {
633         size_t i;
634         int ret;
635
636         for (i = 0; i < ARRAY_SIZE(nxp_c45_hw_stats); i++) {
637                 ret = phy_read_mmd(phydev, nxp_c45_hw_stats[i].mmd,
638                                    nxp_c45_hw_stats[i].reg);
639                 if (ret < 0) {
640                         data[i] = U64_MAX;
641                 } else {
642                         data[i] = ret & nxp_c45_hw_stats[i].mask;
643                         data[i] >>= nxp_c45_hw_stats[i].off;
644                 }
645         }
646 }
647
648 static int nxp_c45_config_enable(struct phy_device *phydev)
649 {
650         phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_DEVICE_CONTROL,
651                       DEVICE_CONTROL_CONFIG_GLOBAL_EN |
652                       DEVICE_CONTROL_CONFIG_ALL_EN);
653         usleep_range(400, 450);
654
655         phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_CONTROL,
656                       PORT_CONTROL_EN);
657         phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONTROL,
658                       PHY_CONFIG_EN);
659         phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_INFRA_CONTROL,
660                       PORT_INFRA_CONTROL_EN);
661
662         return 0;
663 }
664
665 static int nxp_c45_start_op(struct phy_device *phydev)
666 {
667         return phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONTROL,
668                                 PHY_START_OP);
669 }
670
671 static int nxp_c45_config_intr(struct phy_device *phydev)
672 {
673         if (phydev->interrupts == PHY_INTERRUPT_ENABLED)
674                 return phy_set_bits_mmd(phydev, MDIO_MMD_VEND1,
675                                         VEND1_PHY_IRQ_EN, PHY_IRQ_LINK_EVENT);
676         else
677                 return phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1,
678                                           VEND1_PHY_IRQ_EN, PHY_IRQ_LINK_EVENT);
679 }
680
681 static irqreturn_t nxp_c45_handle_interrupt(struct phy_device *phydev)
682 {
683         struct nxp_c45_phy *priv = phydev->priv;
684         irqreturn_t ret = IRQ_NONE;
685         struct nxp_c45_hwts hwts;
686         int irq;
687
688         irq = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_IRQ_STATUS);
689         if (irq & PHY_IRQ_LINK_EVENT) {
690                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_IRQ_ACK,
691                               PHY_IRQ_LINK_EVENT);
692                 phy_trigger_machine(phydev);
693                 ret = IRQ_HANDLED;
694         }
695
696         /* There is no need for ACK.
697          * The irq signal will be asserted until the EGR TS FIFO will be
698          * emptied.
699          */
700         irq = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_PTP_IRQ_STATUS);
701         if (irq & PTP_IRQ_EGR_TS) {
702                 while (nxp_c45_get_hwtxts(priv, &hwts))
703                         nxp_c45_process_txts(priv, &hwts);
704
705                 ret = IRQ_HANDLED;
706         }
707
708         return ret;
709 }
710
711 static int nxp_c45_soft_reset(struct phy_device *phydev)
712 {
713         int ret;
714
715         ret = phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_DEVICE_CONTROL,
716                             DEVICE_CONTROL_RESET);
717         if (ret)
718                 return ret;
719
720         return phy_read_mmd_poll_timeout(phydev, MDIO_MMD_VEND1,
721                                          VEND1_DEVICE_CONTROL, ret,
722                                          !(ret & DEVICE_CONTROL_RESET), 20000,
723                                          240000, false);
724 }
725
726 static int nxp_c45_cable_test_start(struct phy_device *phydev)
727 {
728         return phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_CABLE_TEST,
729                              CABLE_TEST_ENABLE | CABLE_TEST_START);
730 }
731
732 static int nxp_c45_cable_test_get_status(struct phy_device *phydev,
733                                          bool *finished)
734 {
735         int ret;
736         u8 cable_test_result;
737
738         ret = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_CABLE_TEST);
739         if (!(ret & CABLE_TEST_VALID)) {
740                 *finished = false;
741                 return 0;
742         }
743
744         *finished = true;
745         cable_test_result = ret & GENMASK(2, 0);
746
747         switch (cable_test_result) {
748         case CABLE_TEST_OK:
749                 ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
750                                         ETHTOOL_A_CABLE_RESULT_CODE_OK);
751                 break;
752         case CABLE_TEST_SHORTED:
753                 ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
754                                         ETHTOOL_A_CABLE_RESULT_CODE_SAME_SHORT);
755                 break;
756         case CABLE_TEST_OPEN:
757                 ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
758                                         ETHTOOL_A_CABLE_RESULT_CODE_OPEN);
759                 break;
760         default:
761                 ethnl_cable_test_result(phydev, ETHTOOL_A_CABLE_PAIR_A,
762                                         ETHTOOL_A_CABLE_RESULT_CODE_UNSPEC);
763         }
764
765         phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_CABLE_TEST,
766                            CABLE_TEST_ENABLE);
767
768         return nxp_c45_start_op(phydev);
769 }
770
771 static int nxp_c45_setup_master_slave(struct phy_device *phydev)
772 {
773         switch (phydev->master_slave_set) {
774         case MASTER_SLAVE_CFG_MASTER_FORCE:
775         case MASTER_SLAVE_CFG_MASTER_PREFERRED:
776                 phy_write_mmd(phydev, MDIO_MMD_PMAPMD, PMAPMD_B100T1_PMAPMD_CTL,
777                               MASTER_MODE);
778                 break;
779         case MASTER_SLAVE_CFG_SLAVE_PREFERRED:
780         case MASTER_SLAVE_CFG_SLAVE_FORCE:
781                 phy_write_mmd(phydev, MDIO_MMD_PMAPMD, PMAPMD_B100T1_PMAPMD_CTL,
782                               SLAVE_MODE);
783                 break;
784         case MASTER_SLAVE_CFG_UNKNOWN:
785         case MASTER_SLAVE_CFG_UNSUPPORTED:
786                 return 0;
787         default:
788                 phydev_warn(phydev, "Unsupported Master/Slave mode\n");
789                 return -EOPNOTSUPP;
790         }
791
792         return 0;
793 }
794
795 static int nxp_c45_read_master_slave(struct phy_device *phydev)
796 {
797         int reg;
798
799         phydev->master_slave_get = MASTER_SLAVE_CFG_UNKNOWN;
800         phydev->master_slave_state = MASTER_SLAVE_STATE_UNKNOWN;
801
802         reg = phy_read_mmd(phydev, MDIO_MMD_PMAPMD, PMAPMD_B100T1_PMAPMD_CTL);
803         if (reg < 0)
804                 return reg;
805
806         if (reg & B100T1_PMAPMD_MASTER) {
807                 phydev->master_slave_get = MASTER_SLAVE_CFG_MASTER_FORCE;
808                 phydev->master_slave_state = MASTER_SLAVE_STATE_MASTER;
809         } else {
810                 phydev->master_slave_get = MASTER_SLAVE_CFG_SLAVE_FORCE;
811                 phydev->master_slave_state = MASTER_SLAVE_STATE_SLAVE;
812         }
813
814         return 0;
815 }
816
817 static int nxp_c45_config_aneg(struct phy_device *phydev)
818 {
819         return nxp_c45_setup_master_slave(phydev);
820 }
821
822 static int nxp_c45_read_status(struct phy_device *phydev)
823 {
824         int ret;
825
826         ret = genphy_c45_read_status(phydev);
827         if (ret)
828                 return ret;
829
830         ret = nxp_c45_read_master_slave(phydev);
831         if (ret)
832                 return ret;
833
834         return 0;
835 }
836
837 static int nxp_c45_get_sqi(struct phy_device *phydev)
838 {
839         int reg;
840
841         reg = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_SIGNAL_QUALITY);
842         if (!(reg & SQI_VALID))
843                 return -EINVAL;
844
845         reg &= SQI_MASK;
846
847         return reg;
848 }
849
850 static int nxp_c45_get_sqi_max(struct phy_device *phydev)
851 {
852         return MAX_SQI;
853 }
854
855 static int nxp_c45_check_delay(struct phy_device *phydev, u32 delay)
856 {
857         if (delay < MIN_ID_PS) {
858                 phydev_err(phydev, "delay value smaller than %u\n", MIN_ID_PS);
859                 return -EINVAL;
860         }
861
862         if (delay > MAX_ID_PS) {
863                 phydev_err(phydev, "delay value higher than %u\n", MAX_ID_PS);
864                 return -EINVAL;
865         }
866
867         return 0;
868 }
869
870 static u64 nxp_c45_get_phase_shift(u64 phase_offset_raw)
871 {
872         /* The delay in degree phase is 73.8 + phase_offset_raw * 0.9.
873          * To avoid floating point operations we'll multiply by 10
874          * and get 1 decimal point precision.
875          */
876         phase_offset_raw *= 10;
877         phase_offset_raw -= 738;
878         return div_u64(phase_offset_raw, 9);
879 }
880
881 static void nxp_c45_disable_delays(struct phy_device *phydev)
882 {
883         phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID, ID_ENABLE);
884         phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID, ID_ENABLE);
885 }
886
887 static void nxp_c45_set_delays(struct phy_device *phydev)
888 {
889         struct nxp_c45_phy *priv = phydev->priv;
890         u64 tx_delay = priv->tx_delay;
891         u64 rx_delay = priv->rx_delay;
892         u64 degree;
893
894         if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
895             phydev->interface == PHY_INTERFACE_MODE_RGMII_TXID) {
896                 degree = div_u64(tx_delay, PS_PER_DEGREE);
897                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID,
898                               ID_ENABLE | nxp_c45_get_phase_shift(degree));
899         } else {
900                 phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TXID,
901                                    ID_ENABLE);
902         }
903
904         if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
905             phydev->interface == PHY_INTERFACE_MODE_RGMII_RXID) {
906                 degree = div_u64(rx_delay, PS_PER_DEGREE);
907                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID,
908                               ID_ENABLE | nxp_c45_get_phase_shift(degree));
909         } else {
910                 phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RXID,
911                                    ID_ENABLE);
912         }
913 }
914
915 static int nxp_c45_get_delays(struct phy_device *phydev)
916 {
917         struct nxp_c45_phy *priv = phydev->priv;
918         int ret;
919
920         if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
921             phydev->interface == PHY_INTERFACE_MODE_RGMII_TXID) {
922                 ret = device_property_read_u32(&phydev->mdio.dev,
923                                                "tx-internal-delay-ps",
924                                                &priv->tx_delay);
925                 if (ret)
926                         priv->tx_delay = DEFAULT_ID_PS;
927
928                 ret = nxp_c45_check_delay(phydev, priv->tx_delay);
929                 if (ret) {
930                         phydev_err(phydev,
931                                    "tx-internal-delay-ps invalid value\n");
932                         return ret;
933                 }
934         }
935
936         if (phydev->interface == PHY_INTERFACE_MODE_RGMII_ID ||
937             phydev->interface == PHY_INTERFACE_MODE_RGMII_RXID) {
938                 ret = device_property_read_u32(&phydev->mdio.dev,
939                                                "rx-internal-delay-ps",
940                                                &priv->rx_delay);
941                 if (ret)
942                         priv->rx_delay = DEFAULT_ID_PS;
943
944                 ret = nxp_c45_check_delay(phydev, priv->rx_delay);
945                 if (ret) {
946                         phydev_err(phydev,
947                                    "rx-internal-delay-ps invalid value\n");
948                         return ret;
949                 }
950         }
951
952         return 0;
953 }
954
955 static int nxp_c45_set_phy_mode(struct phy_device *phydev)
956 {
957         int ret;
958
959         ret = phy_read_mmd(phydev, MDIO_MMD_VEND1, VEND1_ABILITIES);
960         phydev_dbg(phydev, "Clause 45 managed PHY abilities 0x%x\n", ret);
961
962         switch (phydev->interface) {
963         case PHY_INTERFACE_MODE_RGMII:
964                 if (!(ret & RGMII_ABILITY)) {
965                         phydev_err(phydev, "rgmii mode not supported\n");
966                         return -EINVAL;
967                 }
968                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
969                               MII_BASIC_CONFIG_RGMII);
970                 nxp_c45_disable_delays(phydev);
971                 break;
972         case PHY_INTERFACE_MODE_RGMII_ID:
973         case PHY_INTERFACE_MODE_RGMII_TXID:
974         case PHY_INTERFACE_MODE_RGMII_RXID:
975                 if (!(ret & RGMII_ID_ABILITY)) {
976                         phydev_err(phydev, "rgmii-id, rgmii-txid, rgmii-rxid modes are not supported\n");
977                         return -EINVAL;
978                 }
979                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
980                               MII_BASIC_CONFIG_RGMII);
981                 ret = nxp_c45_get_delays(phydev);
982                 if (ret)
983                         return ret;
984
985                 nxp_c45_set_delays(phydev);
986                 break;
987         case PHY_INTERFACE_MODE_MII:
988                 if (!(ret & MII_ABILITY)) {
989                         phydev_err(phydev, "mii mode not supported\n");
990                         return -EINVAL;
991                 }
992                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
993                               MII_BASIC_CONFIG_MII);
994                 break;
995         case PHY_INTERFACE_MODE_REVMII:
996                 if (!(ret & REVMII_ABILITY)) {
997                         phydev_err(phydev, "rev-mii mode not supported\n");
998                         return -EINVAL;
999                 }
1000                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
1001                               MII_BASIC_CONFIG_MII | MII_BASIC_CONFIG_REV);
1002                 break;
1003         case PHY_INTERFACE_MODE_RMII:
1004                 if (!(ret & RMII_ABILITY)) {
1005                         phydev_err(phydev, "rmii mode not supported\n");
1006                         return -EINVAL;
1007                 }
1008                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
1009                               MII_BASIC_CONFIG_RMII);
1010                 break;
1011         case PHY_INTERFACE_MODE_SGMII:
1012                 if (!(ret & SGMII_ABILITY)) {
1013                         phydev_err(phydev, "sgmii mode not supported\n");
1014                         return -EINVAL;
1015                 }
1016                 phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_MII_BASIC_CONFIG,
1017                               MII_BASIC_CONFIG_SGMII);
1018                 break;
1019         case PHY_INTERFACE_MODE_INTERNAL:
1020                 break;
1021         default:
1022                 return -EINVAL;
1023         }
1024
1025         return 0;
1026 }
1027
1028 static int nxp_c45_config_init(struct phy_device *phydev)
1029 {
1030         int ret;
1031
1032         ret = nxp_c45_config_enable(phydev);
1033         if (ret) {
1034                 phydev_err(phydev, "Failed to enable config\n");
1035                 return ret;
1036         }
1037
1038         /* Bug workaround for SJA1110 rev B: enable write access
1039          * to MDIO_MMD_PMAPMD
1040          */
1041         phy_write_mmd(phydev, MDIO_MMD_VEND1, 0x01F8, 1);
1042         phy_write_mmd(phydev, MDIO_MMD_VEND1, 0x01F9, 2);
1043
1044         phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PHY_CONFIG,
1045                          PHY_CONFIG_AUTO);
1046
1047         phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_LINK_DROP_COUNTER,
1048                          COUNTER_EN);
1049         phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_PREAMBLE_COUNT,
1050                          COUNTER_EN);
1051         phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TX_PREAMBLE_COUNT,
1052                          COUNTER_EN);
1053         phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_IPG_LENGTH,
1054                          COUNTER_EN);
1055         phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_TX_IPG_LENGTH,
1056                          COUNTER_EN);
1057
1058         ret = nxp_c45_set_phy_mode(phydev);
1059         if (ret)
1060                 return ret;
1061
1062         phydev->autoneg = AUTONEG_DISABLE;
1063
1064         phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_PTP_CLK_PERIOD,
1065                       PTP_CLK_PERIOD_100BT1);
1066         phy_clear_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_HW_LTC_LOCK_CTRL,
1067                            HW_LTC_LOCK_EN);
1068         phy_write_mmd(phydev, MDIO_MMD_VEND1, VEND1_RX_TS_INSRT_CTRL,
1069                       RX_TS_INSRT_MODE2);
1070         phy_set_bits_mmd(phydev, MDIO_MMD_VEND1, VEND1_PORT_FUNC_ENABLES,
1071                          PTP_ENABLE);
1072
1073         return nxp_c45_start_op(phydev);
1074 }
1075
1076 static int nxp_c45_probe(struct phy_device *phydev)
1077 {
1078         struct nxp_c45_phy *priv;
1079         int ptp_ability;
1080         int ret = 0;
1081
1082         priv = devm_kzalloc(&phydev->mdio.dev, sizeof(*priv), GFP_KERNEL);
1083         if (!priv)
1084                 return -ENOMEM;
1085
1086         skb_queue_head_init(&priv->tx_queue);
1087         skb_queue_head_init(&priv->rx_queue);
1088
1089         priv->phydev = phydev;
1090
1091         phydev->priv = priv;
1092
1093         mutex_init(&priv->ptp_lock);
1094
1095         ptp_ability = phy_read_mmd(phydev, MDIO_MMD_VEND1,
1096                                    VEND1_PORT_ABILITIES);
1097         ptp_ability = !!(ptp_ability & PTP_ABILITY);
1098         if (!ptp_ability) {
1099                 phydev_dbg(phydev, "the phy does not support PTP");
1100                 goto no_ptp_support;
1101         }
1102
1103         if (IS_ENABLED(CONFIG_PTP_1588_CLOCK) &&
1104             IS_ENABLED(CONFIG_NETWORK_PHY_TIMESTAMPING)) {
1105                 priv->mii_ts.rxtstamp = nxp_c45_rxtstamp;
1106                 priv->mii_ts.txtstamp = nxp_c45_txtstamp;
1107                 priv->mii_ts.hwtstamp = nxp_c45_hwtstamp;
1108                 priv->mii_ts.ts_info = nxp_c45_ts_info;
1109                 phydev->mii_ts = &priv->mii_ts;
1110                 ret = nxp_c45_init_ptp_clock(priv);
1111         } else {
1112                 phydev_dbg(phydev, "PTP support not enabled even if the phy supports it");
1113         }
1114
1115 no_ptp_support:
1116
1117         return ret;
1118 }
1119
1120 static struct phy_driver nxp_c45_driver[] = {
1121         {
1122                 PHY_ID_MATCH_MODEL(PHY_ID_TJA_1103),
1123                 .name                   = "NXP C45 TJA1103",
1124                 .features               = PHY_BASIC_T1_FEATURES,
1125                 .probe                  = nxp_c45_probe,
1126                 .soft_reset             = nxp_c45_soft_reset,
1127                 .config_aneg            = nxp_c45_config_aneg,
1128                 .config_init            = nxp_c45_config_init,
1129                 .config_intr            = nxp_c45_config_intr,
1130                 .handle_interrupt       = nxp_c45_handle_interrupt,
1131                 .read_status            = nxp_c45_read_status,
1132                 .suspend                = genphy_c45_pma_suspend,
1133                 .resume                 = genphy_c45_pma_resume,
1134                 .get_sset_count         = nxp_c45_get_sset_count,
1135                 .get_strings            = nxp_c45_get_strings,
1136                 .get_stats              = nxp_c45_get_stats,
1137                 .cable_test_start       = nxp_c45_cable_test_start,
1138                 .cable_test_get_status  = nxp_c45_cable_test_get_status,
1139                 .set_loopback           = genphy_c45_loopback,
1140                 .get_sqi                = nxp_c45_get_sqi,
1141                 .get_sqi_max            = nxp_c45_get_sqi_max,
1142         },
1143 };
1144
1145 module_phy_driver(nxp_c45_driver);
1146
1147 static struct mdio_device_id __maybe_unused nxp_c45_tbl[] = {
1148         { PHY_ID_MATCH_MODEL(PHY_ID_TJA_1103) },
1149         { /*sentinel*/ },
1150 };
1151
1152 MODULE_DEVICE_TABLE(mdio, nxp_c45_tbl);
1153
1154 MODULE_AUTHOR("Radu Pirea <radu-nicolae.pirea@oss.nxp.com>");
1155 MODULE_DESCRIPTION("NXP C45 PHY driver");
1156 MODULE_LICENSE("GPL v2");