1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3 * Copyright 2015-2017 Google, Inc
4 *
5 * USB Power Delivery protocol stack.
6 */
7
8 #include <linux/completion.h>
9 #include <linux/debugfs.h>
10 #include <linux/device.h>
11 #include <linux/hrtimer.h>
12 #include <linux/jiffies.h>
13 #include <linux/kernel.h>
14 #include <linux/kthread.h>
15 #include <linux/module.h>
16 #include <linux/mutex.h>
17 #include <linux/power_supply.h>
18 #include <linux/proc_fs.h>
19 #include <linux/property.h>
20 #include <linux/sched/clock.h>
21 #include <linux/seq_file.h>
22 #include <linux/slab.h>
23 #include <linux/spinlock.h>
24 #include <linux/string_choices.h>
25 #include <linux/usb.h>
26 #include <linux/usb/pd.h>
27 #include <linux/usb/pd_ado.h>
28 #include <linux/usb/pd_bdo.h>
29 #include <linux/usb/pd_ext_sdb.h>
30 #include <linux/usb/pd_vdo.h>
31 #include <linux/usb/role.h>
32 #include <linux/usb/tcpm.h>
33 #include <linux/usb/typec_altmode.h>
34
35 #include <uapi/linux/sched/types.h>
36
37 #define FOREACH_STATE(S) \
38 S(INVALID_STATE), \
39 S(TOGGLING), \
40 S(CHECK_CONTAMINANT), \
41 S(SRC_UNATTACHED), \
42 S(SRC_ATTACH_WAIT), \
43 S(SRC_ATTACHED), \
44 S(SRC_STARTUP), \
45 S(SRC_SEND_CAPABILITIES), \
46 S(SRC_SEND_CAPABILITIES_TIMEOUT), \
47 S(SRC_NEGOTIATE_CAPABILITIES), \
48 S(SRC_TRANSITION_SUPPLY), \
49 S(SRC_READY), \
50 S(SRC_WAIT_NEW_CAPABILITIES), \
51 \
52 S(SNK_UNATTACHED), \
53 S(SNK_ATTACH_WAIT), \
54 S(SNK_DEBOUNCED), \
55 S(SNK_ATTACHED), \
56 S(SNK_STARTUP), \
57 S(SNK_DISCOVERY), \
58 S(SNK_DISCOVERY_DEBOUNCE), \
59 S(SNK_DISCOVERY_DEBOUNCE_DONE), \
60 S(SNK_WAIT_CAPABILITIES), \
61 S(SNK_WAIT_CAPABILITIES_TIMEOUT), \
62 S(SNK_NEGOTIATE_CAPABILITIES), \
63 S(SNK_NEGOTIATE_PPS_CAPABILITIES), \
64 S(SNK_TRANSITION_SINK), \
65 S(SNK_TRANSITION_SINK_VBUS), \
66 S(SNK_READY), \
67 \
68 S(ACC_UNATTACHED), \
69 S(DEBUG_ACC_ATTACHED), \
70 S(DEBUG_ACC_DEBOUNCE), \
71 S(AUDIO_ACC_ATTACHED), \
72 S(AUDIO_ACC_DEBOUNCE), \
73 \
74 S(HARD_RESET_SEND), \
75 S(HARD_RESET_START), \
76 S(SRC_HARD_RESET_VBUS_OFF), \
77 S(SRC_HARD_RESET_VBUS_ON), \
78 S(SNK_HARD_RESET_SINK_OFF), \
79 S(SNK_HARD_RESET_WAIT_VBUS), \
80 S(SNK_HARD_RESET_SINK_ON), \
81 \
82 S(SOFT_RESET), \
83 S(SRC_SOFT_RESET_WAIT_SNK_TX), \
84 S(SNK_SOFT_RESET), \
85 S(SOFT_RESET_SEND), \
86 \
87 S(DR_SWAP_ACCEPT), \
88 S(DR_SWAP_SEND), \
89 S(DR_SWAP_SEND_TIMEOUT), \
90 S(DR_SWAP_CANCEL), \
91 S(DR_SWAP_CHANGE_DR), \
92 \
93 S(PR_SWAP_ACCEPT), \
94 S(PR_SWAP_SEND), \
95 S(PR_SWAP_SEND_TIMEOUT), \
96 S(PR_SWAP_CANCEL), \
97 S(PR_SWAP_START), \
98 S(PR_SWAP_SRC_SNK_TRANSITION_OFF), \
99 S(PR_SWAP_SRC_SNK_SOURCE_OFF), \
100 S(PR_SWAP_SRC_SNK_SOURCE_OFF_CC_DEBOUNCED), \
101 S(PR_SWAP_SRC_SNK_SINK_ON), \
102 S(PR_SWAP_SNK_SRC_SINK_OFF), \
103 S(PR_SWAP_SNK_SRC_SOURCE_ON), \
104 S(PR_SWAP_SNK_SRC_SOURCE_ON_VBUS_RAMPED_UP), \
105 \
106 S(VCONN_SWAP_ACCEPT), \
107 S(VCONN_SWAP_SEND), \
108 S(VCONN_SWAP_SEND_TIMEOUT), \
109 S(VCONN_SWAP_CANCEL), \
110 S(VCONN_SWAP_START), \
111 S(VCONN_SWAP_WAIT_FOR_VCONN), \
112 S(VCONN_SWAP_TURN_ON_VCONN), \
113 S(VCONN_SWAP_TURN_OFF_VCONN), \
114 S(VCONN_SWAP_SEND_SOFT_RESET), \
115 \
116 S(FR_SWAP_SEND), \
117 S(FR_SWAP_SEND_TIMEOUT), \
118 S(FR_SWAP_SNK_SRC_TRANSITION_TO_OFF), \
119 S(FR_SWAP_SNK_SRC_NEW_SINK_READY), \
120 S(FR_SWAP_SNK_SRC_SOURCE_VBUS_APPLIED), \
121 S(FR_SWAP_CANCEL), \
122 \
123 S(SNK_TRY), \
124 S(SNK_TRY_WAIT), \
125 S(SNK_TRY_WAIT_DEBOUNCE), \
126 S(SNK_TRY_WAIT_DEBOUNCE_CHECK_VBUS), \
127 S(SRC_TRYWAIT), \
128 S(SRC_TRYWAIT_DEBOUNCE), \
129 S(SRC_TRYWAIT_UNATTACHED), \
130 \
131 S(SRC_TRY), \
132 S(SRC_TRY_WAIT), \
133 S(SRC_TRY_DEBOUNCE), \
134 S(SNK_TRYWAIT), \
135 S(SNK_TRYWAIT_DEBOUNCE), \
136 S(SNK_TRYWAIT_VBUS), \
137 S(BIST_RX), \
138 \
139 S(GET_STATUS_SEND), \
140 S(GET_STATUS_SEND_TIMEOUT), \
141 S(GET_PPS_STATUS_SEND), \
142 S(GET_PPS_STATUS_SEND_TIMEOUT), \
143 \
144 S(GET_SINK_CAP), \
145 S(GET_SINK_CAP_TIMEOUT), \
146 \
147 S(ERROR_RECOVERY), \
148 S(PORT_RESET), \
149 S(PORT_RESET_WAIT_OFF), \
150 \
151 S(AMS_START), \
152 S(CHUNK_NOT_SUPP), \
153 \
154 S(SRC_VDM_IDENTITY_REQUEST)
155
156 #define FOREACH_AMS(S) \
157 S(NONE_AMS), \
158 S(POWER_NEGOTIATION), \
159 S(GOTOMIN), \
160 S(SOFT_RESET_AMS), \
161 S(HARD_RESET), \
162 S(CABLE_RESET), \
163 S(GET_SOURCE_CAPABILITIES), \
164 S(GET_SINK_CAPABILITIES), \
165 S(POWER_ROLE_SWAP), \
166 S(FAST_ROLE_SWAP), \
167 S(DATA_ROLE_SWAP), \
168 S(VCONN_SWAP), \
169 S(SOURCE_ALERT), \
170 S(GETTING_SOURCE_EXTENDED_CAPABILITIES),\
171 S(GETTING_SOURCE_SINK_STATUS), \
172 S(GETTING_BATTERY_CAPABILITIES), \
173 S(GETTING_BATTERY_STATUS), \
174 S(GETTING_MANUFACTURER_INFORMATION), \
175 S(SECURITY), \
176 S(FIRMWARE_UPDATE), \
177 S(DISCOVER_IDENTITY), \
178 S(SOURCE_STARTUP_CABLE_PLUG_DISCOVER_IDENTITY), \
179 S(DISCOVER_SVIDS), \
180 S(DISCOVER_MODES), \
181 S(DFP_TO_UFP_ENTER_MODE), \
182 S(DFP_TO_UFP_EXIT_MODE), \
183 S(DFP_TO_CABLE_PLUG_ENTER_MODE), \
184 S(DFP_TO_CABLE_PLUG_EXIT_MODE), \
185 S(ATTENTION), \
186 S(BIST), \
187 S(UNSTRUCTURED_VDMS), \
188 S(STRUCTURED_VDMS), \
189 S(COUNTRY_INFO), \
190 S(COUNTRY_CODES), \
191 S(REVISION_INFORMATION)
192
193 #define GENERATE_ENUM(e) e
194 #define GENERATE_STRING(s) #s
195
196 enum tcpm_state {
197 FOREACH_STATE(GENERATE_ENUM)
198 };
199
200 static const char * const tcpm_states[] = {
201 FOREACH_STATE(GENERATE_STRING)
202 };
203
204 enum tcpm_ams {
205 FOREACH_AMS(GENERATE_ENUM)
206 };
207
208 static const char * const tcpm_ams_str[] = {
209 FOREACH_AMS(GENERATE_STRING)
210 };
211
212 enum vdm_states {
213 VDM_STATE_ERR_BUSY = -3,
214 VDM_STATE_ERR_SEND = -2,
215 VDM_STATE_ERR_TMOUT = -1,
216 VDM_STATE_DONE = 0,
217 /* Anything >0 represents an active state */
218 VDM_STATE_READY = 1,
219 VDM_STATE_BUSY = 2,
220 VDM_STATE_WAIT_RSP_BUSY = 3,
221 VDM_STATE_SEND_MESSAGE = 4,
222 };
223
224 enum pd_msg_request {
225 PD_MSG_NONE = 0,
226 PD_MSG_CTRL_REJECT,
227 PD_MSG_CTRL_WAIT,
228 PD_MSG_CTRL_NOT_SUPP,
229 PD_MSG_DATA_SINK_CAP,
230 PD_MSG_DATA_SOURCE_CAP,
231 PD_MSG_DATA_REV,
232 };
233
234 enum adev_actions {
235 ADEV_NONE = 0,
236 ADEV_NOTIFY_USB_AND_QUEUE_VDM,
237 ADEV_QUEUE_VDM,
238 ADEV_QUEUE_VDM_SEND_EXIT_MODE_ON_FAIL,
239 ADEV_ATTENTION,
240 };
241
242 /*
243 * Initial current capability of the new source when vSafe5V is applied during PD3.0 Fast Role Swap.
244 * Based on "Table 6-14 Fixed Supply PDO - Sink" of "USB Power Delivery Specification Revision 3.0,
245 * Version 1.2"
246 */
247 enum frs_typec_current {
248 FRS_NOT_SUPPORTED,
249 FRS_DEFAULT_POWER,
250 FRS_5V_1P5A,
251 FRS_5V_3A,
252 };
253
254 /* Events from low level driver */
255
256 #define TCPM_CC_EVENT BIT(0)
257 #define TCPM_VBUS_EVENT BIT(1)
258 #define TCPM_RESET_EVENT BIT(2)
259 #define TCPM_FRS_EVENT BIT(3)
260 #define TCPM_SOURCING_VBUS BIT(4)
261 #define TCPM_PORT_CLEAN BIT(5)
262 #define TCPM_PORT_ERROR BIT(6)
263
264 #define LOG_BUFFER_ENTRIES 1024
265 #define LOG_BUFFER_ENTRY_SIZE 128
266
267 /* Alternate mode support */
268
269 #define SVID_DISCOVERY_MAX 16
270 #define ALTMODE_DISCOVERY_MAX (SVID_DISCOVERY_MAX * MODE_DISCOVERY_MAX)
271
272 #define GET_SINK_CAP_RETRY_MS 100
273 #define SEND_DISCOVER_RETRY_MS 100
274
275 struct pd_mode_data {
276 int svid_index; /* current SVID index */
277 int nsvids;
278 u16 svids[SVID_DISCOVERY_MAX];
279 int altmodes; /* number of alternate modes */
280 struct typec_altmode_desc altmode_desc[ALTMODE_DISCOVERY_MAX];
281 };
282
283 /*
284 * @min_volt: Actual min voltage at the local port
285 * @req_min_volt: Requested min voltage to the port partner
286 * @max_volt: Actual max voltage at the local port
287 * @req_max_volt: Requested max voltage to the port partner
288 * @max_curr: Actual max current at the local port
289 * @req_max_curr: Requested max current of the port partner
290 * @req_out_volt: Requested output voltage to the port partner
291 * @req_op_curr: Requested operating current to the port partner
292 * @supported: Parter has at least one APDO hence supports PPS
293 * @active: PPS mode is active
294 */
295 struct pd_pps_data {
296 u32 min_volt;
297 u32 req_min_volt;
298 u32 max_volt;
299 u32 req_max_volt;
300 u32 max_curr;
301 u32 req_max_curr;
302 u32 req_out_volt;
303 u32 req_op_curr;
304 bool supported;
305 bool active;
306 };
307
308 struct pd_data {
309 struct usb_power_delivery *pd;
310 struct usb_power_delivery_capabilities *source_cap;
311 struct usb_power_delivery_capabilities_desc source_desc;
312 struct usb_power_delivery_capabilities *sink_cap;
313 struct usb_power_delivery_capabilities_desc sink_desc;
314 unsigned int operating_snk_mw;
315 };
316
317 #define PD_CAP_REV10 0x1
318 #define PD_CAP_REV20 0x2
319 #define PD_CAP_REV30 0x3
320
321 struct pd_revision_info {
322 u8 rev_major;
323 u8 rev_minor;
324 u8 ver_major;
325 u8 ver_minor;
326 };
327
328 /*
329 * @sink_wait_cap_time: Deadline (in ms) for tTypeCSinkWaitCap timer
330 * @ps_src_wait_off_time: Deadline (in ms) for tPSSourceOff timer
331 * @cc_debounce_time: Deadline (in ms) for tCCDebounce timer
332 */
333 struct pd_timings {
334 u32 sink_wait_cap_time;
335 u32 ps_src_off_time;
336 u32 cc_debounce_time;
337 u32 snk_bc12_cmpletion_time;
338 };
339
340 struct tcpm_port {
341 struct device *dev;
342
343 struct mutex lock; /* tcpm state machine lock */
344 struct kthread_worker *wq;
345
346 struct typec_capability typec_caps;
347 struct typec_port *typec_port;
348
349 struct tcpc_dev *tcpc;
350 struct usb_role_switch *role_sw;
351
352 enum typec_role vconn_role;
353 enum typec_role pwr_role;
354 enum typec_data_role data_role;
355 enum typec_pwr_opmode pwr_opmode;
356
357 struct usb_pd_identity partner_ident;
358 struct typec_partner_desc partner_desc;
359 struct typec_partner *partner;
360
361 struct usb_pd_identity cable_ident;
362 struct typec_cable_desc cable_desc;
363 struct typec_cable *cable;
364 struct typec_plug_desc plug_prime_desc;
365 struct typec_plug *plug_prime;
366
367 enum typec_cc_status cc_req;
368 enum typec_cc_status src_rp; /* work only if pd_supported == false */
369
370 enum typec_cc_status cc1;
371 enum typec_cc_status cc2;
372 enum typec_cc_polarity polarity;
373
374 bool attached;
375 bool connected;
376 bool registered;
377 bool pd_supported;
378 enum typec_port_type port_type;
379
380 /*
381 * Set to true when vbus is greater than VSAFE5V min.
382 * Set to false when vbus falls below vSinkDisconnect max threshold.
383 */
384 bool vbus_present;
385
386 /*
387 * Set to true when vbus is less than VSAFE0V max.
388 * Set to false when vbus is greater than VSAFE0V max.
389 */
390 bool vbus_vsafe0v;
391
392 bool vbus_never_low;
393 bool vbus_source;
394 bool vbus_charge;
395
396 /* Set to true when Discover_Identity Command is expected to be sent in Ready states. */
397 bool send_discover;
398 bool op_vsafe5v;
399
400 int try_role;
401 int try_snk_count;
402 int try_src_count;
403
404 enum pd_msg_request queued_message;
405
406 enum tcpm_state enter_state;
407 enum tcpm_state prev_state;
408 enum tcpm_state state;
409 enum tcpm_state delayed_state;
410 ktime_t delayed_runtime;
411 unsigned long delay_ms;
412
413 spinlock_t pd_event_lock;
414 u32 pd_events;
415
416 struct kthread_work event_work;
417 struct hrtimer state_machine_timer;
418 struct kthread_work state_machine;
419 struct hrtimer vdm_state_machine_timer;
420 struct kthread_work vdm_state_machine;
421 struct hrtimer enable_frs_timer;
422 struct kthread_work enable_frs;
423 struct hrtimer send_discover_timer;
424 struct kthread_work send_discover_work;
425 bool state_machine_running;
426 /* Set to true when VDM State Machine has following actions. */
427 bool vdm_sm_running;
428
429 struct completion tx_complete;
430 enum tcpm_transmit_status tx_status;
431
432 struct mutex swap_lock; /* swap command lock */
433 bool swap_pending;
434 bool non_pd_role_swap;
435 struct completion swap_complete;
436 int swap_status;
437
438 unsigned int negotiated_rev;
439 unsigned int message_id;
440 unsigned int caps_count;
441 unsigned int hard_reset_count;
442 bool pd_capable;
443 bool explicit_contract;
444 unsigned int rx_msgid;
445
446 /* USB PD objects */
447 struct usb_power_delivery **pds;
448 struct pd_data **pd_list;
449 struct usb_power_delivery_capabilities *port_source_caps;
450 struct usb_power_delivery_capabilities *port_sink_caps;
451 struct usb_power_delivery *partner_pd;
452 struct usb_power_delivery_capabilities *partner_source_caps;
453 struct usb_power_delivery_capabilities *partner_sink_caps;
454 struct usb_power_delivery *selected_pd;
455
456 /* Partner capabilities/requests */
457 u32 sink_request;
458 u32 source_caps[PDO_MAX_OBJECTS];
459 unsigned int nr_source_caps;
460 u32 sink_caps[PDO_MAX_OBJECTS];
461 unsigned int nr_sink_caps;
462
463 /* Local capabilities */
464 unsigned int pd_count;
465 u32 src_pdo[PDO_MAX_OBJECTS];
466 unsigned int nr_src_pdo;
467 u32 snk_pdo[PDO_MAX_OBJECTS];
468 unsigned int nr_snk_pdo;
469 u32 snk_vdo_v1[VDO_MAX_OBJECTS];
470 unsigned int nr_snk_vdo_v1;
471 u32 snk_vdo[VDO_MAX_OBJECTS];
472 unsigned int nr_snk_vdo;
473
474 unsigned int operating_snk_mw;
475 bool update_sink_caps;
476
477 /* Requested current / voltage to the port partner */
478 u32 req_current_limit;
479 u32 req_supply_voltage;
480 /* Actual current / voltage limit of the local port */
481 u32 current_limit;
482 u32 supply_voltage;
483
484 /* Used to export TA voltage and current */
485 struct power_supply *psy;
486 struct power_supply_desc psy_desc;
487 enum power_supply_usb_type usb_type;
488
489 u32 bist_request;
490
491 /* PD state for Vendor Defined Messages */
492 enum vdm_states vdm_state;
493 u32 vdm_retries;
494 /* next Vendor Defined Message to send */
495 u32 vdo_data[VDO_MAX_SIZE];
496 u8 vdo_count;
497 /* VDO to retry if UFP responder replied busy */
498 u32 vdo_retry;
499
500 /* PPS */
501 struct pd_pps_data pps_data;
502 struct completion pps_complete;
503 bool pps_pending;
504 int pps_status;
505
506 /* Alternate mode data */
507 struct pd_mode_data mode_data;
508 struct pd_mode_data mode_data_prime;
509 struct typec_altmode *partner_altmode[ALTMODE_DISCOVERY_MAX];
510 struct typec_altmode *plug_prime_altmode[ALTMODE_DISCOVERY_MAX];
511 struct typec_altmode *port_altmode[ALTMODE_DISCOVERY_MAX];
512
513 /* Deadline in jiffies to exit src_try_wait state */
514 unsigned long max_wait;
515
516 /* port belongs to a self powered device */
517 bool self_powered;
518
519 /* Sink FRS */
520 enum frs_typec_current new_source_frs_current;
521
522 /* Sink caps have been queried */
523 bool sink_cap_done;
524
525 /* Collision Avoidance and Atomic Message Sequence */
526 enum tcpm_state upcoming_state;
527 enum tcpm_ams ams;
528 enum tcpm_ams next_ams;
529 bool in_ams;
530
531 /* Auto vbus discharge status */
532 bool auto_vbus_discharge_enabled;
533
534 /*
535 * When set, port requests PD_P_SNK_STDBY_MW upon entering SNK_DISCOVERY and
536 * the actual current limit after RX of PD_CTRL_PSRDY for PD link,
537 * SNK_READY for non-pd link.
538 */
539 bool slow_charger_loop;
540
541 /*
542 * When true indicates that the lower level drivers indicate potential presence
543 * of contaminant in the connector pins based on the tcpm state machine
544 * transitions.
545 */
546 bool potential_contaminant;
547
548 /* SOP* Related Fields */
549 /*
550 * Flag to determine if SOP' Discover Identity is available. The flag
551 * is set if Discover Identity on SOP' does not immediately follow
552 * Discover Identity on SOP.
553 */
554 bool send_discover_prime;
555 /*
556 * tx_sop_type determines which SOP* a message is being sent on.
557 * For messages that are queued and not sent immediately such as in
558 * tcpm_queue_message or messages that send after state changes,
559 * the tx_sop_type is set accordingly.
560 */
561 enum tcpm_transmit_type tx_sop_type;
562 /*
563 * Prior to discovering the port partner's Specification Revision, the
564 * Vconn source and cable plug will use the lower of their two revisions.
565 *
566 * When the port partner's Specification Revision is discovered, the following
567 * rules are put in place.
568 * 1. If the cable revision (1) is lower than the revision negotiated
569 * between the port and partner (2), the port and partner will communicate
570 * on revision (2), but the port and cable will communicate on revision (1).
571 * 2. If the cable revision (1) is higher than the revision negotiated
572 * between the port and partner (2), the port and partner will communicate
573 * on revision (2), and the port and cable will communicate on revision (2)
574 * as well.
575 */
576 unsigned int negotiated_rev_prime;
577 /*
578 * Each SOP* type must maintain their own tx and rx message IDs
579 */
580 unsigned int message_id_prime;
581 unsigned int rx_msgid_prime;
582
583 /* Timer deadline values configured at runtime */
584 struct pd_timings timings;
585
586 /* Indicates maximum (revision, version) supported */
587 struct pd_revision_info pd_rev;
588 #ifdef CONFIG_DEBUG_FS
589 struct dentry *dentry;
590 struct mutex logbuffer_lock; /* log buffer access lock */
591 int logbuffer_head;
592 int logbuffer_tail;
593 u8 *logbuffer[LOG_BUFFER_ENTRIES];
594 #endif
595 };
596
597 struct pd_rx_event {
598 struct kthread_work work;
599 struct tcpm_port *port;
600 struct pd_message msg;
601 enum tcpm_transmit_type rx_sop_type;
602 };
603
604 struct altmode_vdm_event {
605 struct kthread_work work;
606 struct tcpm_port *port;
607 u32 header;
608 u32 *data;
609 int cnt;
610 enum tcpm_transmit_type tx_sop_type;
611 };
612
613 static const char * const pd_rev[] = {
614 [PD_REV10] = "rev1",
615 [PD_REV20] = "rev2",
616 [PD_REV30] = "rev3",
617 };
618
619 #define tcpm_cc_is_sink(cc) \
620 ((cc) == TYPEC_CC_RP_DEF || (cc) == TYPEC_CC_RP_1_5 || \
621 (cc) == TYPEC_CC_RP_3_0)
622
623 /* As long as cc is pulled up, we can consider it as sink. */
624 #define tcpm_port_is_sink(port) \
625 (tcpm_cc_is_sink((port)->cc1) || tcpm_cc_is_sink((port)->cc2))
626
627 #define tcpm_cc_is_source(cc) ((cc) == TYPEC_CC_RD)
628 #define tcpm_cc_is_audio(cc) ((cc) == TYPEC_CC_RA)
629 #define tcpm_cc_is_open(cc) ((cc) == TYPEC_CC_OPEN)
630
631 #define tcpm_port_is_source(port) \
632 ((tcpm_cc_is_source((port)->cc1) && \
633 !tcpm_cc_is_source((port)->cc2)) || \
634 (tcpm_cc_is_source((port)->cc2) && \
635 !tcpm_cc_is_source((port)->cc1)))
636
637 #define tcpm_port_is_debug(port) \
638 ((tcpm_cc_is_source((port)->cc1) && tcpm_cc_is_source((port)->cc2)) || \
639 (tcpm_cc_is_sink((port)->cc1) && tcpm_cc_is_sink((port)->cc2)))
640
641 #define tcpm_port_is_audio(port) \
642 (tcpm_cc_is_audio((port)->cc1) && tcpm_cc_is_audio((port)->cc2))
643
644 #define tcpm_port_is_audio_detached(port) \
645 ((tcpm_cc_is_audio((port)->cc1) && tcpm_cc_is_open((port)->cc2)) || \
646 (tcpm_cc_is_audio((port)->cc2) && tcpm_cc_is_open((port)->cc1)))
647
648 #define tcpm_try_snk(port) \
649 ((port)->try_snk_count == 0 && (port)->try_role == TYPEC_SINK && \
650 (port)->port_type == TYPEC_PORT_DRP)
651
652 #define tcpm_try_src(port) \
653 ((port)->try_src_count == 0 && (port)->try_role == TYPEC_SOURCE && \
654 (port)->port_type == TYPEC_PORT_DRP)
655
656 #define tcpm_data_role_for_source(port) \
657 ((port)->typec_caps.data == TYPEC_PORT_UFP ? \
658 TYPEC_DEVICE : TYPEC_HOST)
659
660 #define tcpm_data_role_for_sink(port) \
661 ((port)->typec_caps.data == TYPEC_PORT_DFP ? \
662 TYPEC_HOST : TYPEC_DEVICE)
663
664 #define tcpm_sink_tx_ok(port) \
665 (tcpm_port_is_sink(port) && \
666 ((port)->cc1 == TYPEC_CC_RP_3_0 || (port)->cc2 == TYPEC_CC_RP_3_0))
667
668 #define tcpm_wait_for_discharge(port) \
669 (((port)->auto_vbus_discharge_enabled && !(port)->vbus_vsafe0v) ? PD_T_SAFE_0V : 0)
670
tcpm_default_state(struct tcpm_port * port)671 static enum tcpm_state tcpm_default_state(struct tcpm_port *port)
672 {
673 if (port->port_type == TYPEC_PORT_DRP) {
674 if (port->try_role == TYPEC_SINK)
675 return SNK_UNATTACHED;
676 else if (port->try_role == TYPEC_SOURCE)
677 return SRC_UNATTACHED;
678 /* Fall through to return SRC_UNATTACHED */
679 } else if (port->port_type == TYPEC_PORT_SNK) {
680 return SNK_UNATTACHED;
681 }
682 return SRC_UNATTACHED;
683 }
684
tcpm_port_is_disconnected(struct tcpm_port * port)685 static bool tcpm_port_is_disconnected(struct tcpm_port *port)
686 {
687 return (!port->attached && port->cc1 == TYPEC_CC_OPEN &&
688 port->cc2 == TYPEC_CC_OPEN) ||
689 (port->attached && ((port->polarity == TYPEC_POLARITY_CC1 &&
690 port->cc1 == TYPEC_CC_OPEN) ||
691 (port->polarity == TYPEC_POLARITY_CC2 &&
692 port->cc2 == TYPEC_CC_OPEN)));
693 }
694
695 /*
696 * Logging
697 */
698
699 #ifdef CONFIG_DEBUG_FS
700
tcpm_log_full(struct tcpm_port * port)701 static bool tcpm_log_full(struct tcpm_port *port)
702 {
703 return port->logbuffer_tail ==
704 (port->logbuffer_head + 1) % LOG_BUFFER_ENTRIES;
705 }
706
707 __printf(2, 0)
_tcpm_log(struct tcpm_port * port,const char * fmt,va_list args)708 static void _tcpm_log(struct tcpm_port *port, const char *fmt, va_list args)
709 {
710 char tmpbuffer[LOG_BUFFER_ENTRY_SIZE];
711 u64 ts_nsec = local_clock();
712 unsigned long rem_nsec;
713
714 mutex_lock(&port->logbuffer_lock);
715 if (!port->logbuffer[port->logbuffer_head]) {
716 port->logbuffer[port->logbuffer_head] =
717 kzalloc(LOG_BUFFER_ENTRY_SIZE, GFP_KERNEL);
718 if (!port->logbuffer[port->logbuffer_head]) {
719 mutex_unlock(&port->logbuffer_lock);
720 return;
721 }
722 }
723
724 vsnprintf(tmpbuffer, sizeof(tmpbuffer), fmt, args);
725
726 if (tcpm_log_full(port)) {
727 port->logbuffer_head = max(port->logbuffer_head - 1, 0);
728 strcpy(tmpbuffer, "overflow");
729 }
730
731 if (port->logbuffer_head < 0 ||
732 port->logbuffer_head >= LOG_BUFFER_ENTRIES) {
733 dev_warn(port->dev,
734 "Bad log buffer index %d\n", port->logbuffer_head);
735 goto abort;
736 }
737
738 if (!port->logbuffer[port->logbuffer_head]) {
739 dev_warn(port->dev,
740 "Log buffer index %d is NULL\n", port->logbuffer_head);
741 goto abort;
742 }
743
744 rem_nsec = do_div(ts_nsec, 1000000000);
745 scnprintf(port->logbuffer[port->logbuffer_head],
746 LOG_BUFFER_ENTRY_SIZE, "[%5lu.%06lu] %s",
747 (unsigned long)ts_nsec, rem_nsec / 1000,
748 tmpbuffer);
749 port->logbuffer_head = (port->logbuffer_head + 1) % LOG_BUFFER_ENTRIES;
750
751 abort:
752 mutex_unlock(&port->logbuffer_lock);
753 }
754
755 __printf(2, 3)
tcpm_log(struct tcpm_port * port,const char * fmt,...)756 static void tcpm_log(struct tcpm_port *port, const char *fmt, ...)
757 {
758 va_list args;
759
760 /* Do not log while disconnected and unattached */
761 if (tcpm_port_is_disconnected(port) &&
762 (port->state == SRC_UNATTACHED || port->state == SNK_UNATTACHED ||
763 port->state == TOGGLING || port->state == CHECK_CONTAMINANT))
764 return;
765
766 va_start(args, fmt);
767 _tcpm_log(port, fmt, args);
768 va_end(args);
769 }
770
771 __printf(2, 3)
tcpm_log_force(struct tcpm_port * port,const char * fmt,...)772 static void tcpm_log_force(struct tcpm_port *port, const char *fmt, ...)
773 {
774 va_list args;
775
776 va_start(args, fmt);
777 _tcpm_log(port, fmt, args);
778 va_end(args);
779 }
780
tcpm_log_source_caps(struct tcpm_port * port)781 static void tcpm_log_source_caps(struct tcpm_port *port)
782 {
783 int i;
784
785 for (i = 0; i < port->nr_source_caps; i++) {
786 u32 pdo = port->source_caps[i];
787 enum pd_pdo_type type = pdo_type(pdo);
788 char msg[64];
789
790 switch (type) {
791 case PDO_TYPE_FIXED:
792 scnprintf(msg, sizeof(msg),
793 "%u mV, %u mA [%s%s%s%s%s%s]",
794 pdo_fixed_voltage(pdo),
795 pdo_max_current(pdo),
796 (pdo & PDO_FIXED_DUAL_ROLE) ?
797 "R" : "",
798 (pdo & PDO_FIXED_SUSPEND) ?
799 "S" : "",
800 (pdo & PDO_FIXED_HIGHER_CAP) ?
801 "H" : "",
802 (pdo & PDO_FIXED_USB_COMM) ?
803 "U" : "",
804 (pdo & PDO_FIXED_DATA_SWAP) ?
805 "D" : "",
806 (pdo & PDO_FIXED_EXTPOWER) ?
807 "E" : "");
808 break;
809 case PDO_TYPE_VAR:
810 scnprintf(msg, sizeof(msg),
811 "%u-%u mV, %u mA",
812 pdo_min_voltage(pdo),
813 pdo_max_voltage(pdo),
814 pdo_max_current(pdo));
815 break;
816 case PDO_TYPE_BATT:
817 scnprintf(msg, sizeof(msg),
818 "%u-%u mV, %u mW",
819 pdo_min_voltage(pdo),
820 pdo_max_voltage(pdo),
821 pdo_max_power(pdo));
822 break;
823 case PDO_TYPE_APDO:
824 if (pdo_apdo_type(pdo) == APDO_TYPE_PPS)
825 scnprintf(msg, sizeof(msg),
826 "%u-%u mV, %u mA",
827 pdo_pps_apdo_min_voltage(pdo),
828 pdo_pps_apdo_max_voltage(pdo),
829 pdo_pps_apdo_max_current(pdo));
830 else
831 strcpy(msg, "undefined APDO");
832 break;
833 default:
834 strcpy(msg, "undefined");
835 break;
836 }
837 tcpm_log(port, " PDO %d: type %d, %s",
838 i, type, msg);
839 }
840 }
841
tcpm_debug_show(struct seq_file * s,void * v)842 static int tcpm_debug_show(struct seq_file *s, void *v)
843 {
844 struct tcpm_port *port = s->private;
845 int tail;
846
847 mutex_lock(&port->logbuffer_lock);
848 tail = port->logbuffer_tail;
849 while (tail != port->logbuffer_head) {
850 seq_printf(s, "%s\n", port->logbuffer[tail]);
851 tail = (tail + 1) % LOG_BUFFER_ENTRIES;
852 }
853 if (!seq_has_overflowed(s))
854 port->logbuffer_tail = tail;
855 mutex_unlock(&port->logbuffer_lock);
856
857 return 0;
858 }
859 DEFINE_SHOW_ATTRIBUTE(tcpm_debug);
860
tcpm_debugfs_init(struct tcpm_port * port)861 static void tcpm_debugfs_init(struct tcpm_port *port)
862 {
863 char name[NAME_MAX];
864
865 mutex_init(&port->logbuffer_lock);
866 snprintf(name, NAME_MAX, "tcpm-%s", dev_name(port->dev));
867 port->dentry = debugfs_create_dir(name, usb_debug_root);
868 debugfs_create_file("log", S_IFREG | 0444, port->dentry, port,
869 &tcpm_debug_fops);
870 }
871
tcpm_debugfs_exit(struct tcpm_port * port)872 static void tcpm_debugfs_exit(struct tcpm_port *port)
873 {
874 int i;
875
876 mutex_lock(&port->logbuffer_lock);
877 for (i = 0; i < LOG_BUFFER_ENTRIES; i++) {
878 kfree(port->logbuffer[i]);
879 port->logbuffer[i] = NULL;
880 }
881 mutex_unlock(&port->logbuffer_lock);
882
883 debugfs_remove(port->dentry);
884 }
885
886 #else
887
888 __printf(2, 3)
tcpm_log(const struct tcpm_port * port,const char * fmt,...)889 static void tcpm_log(const struct tcpm_port *port, const char *fmt, ...) { }
890 __printf(2, 3)
tcpm_log_force(struct tcpm_port * port,const char * fmt,...)891 static void tcpm_log_force(struct tcpm_port *port, const char *fmt, ...) { }
tcpm_log_source_caps(struct tcpm_port * port)892 static void tcpm_log_source_caps(struct tcpm_port *port) { }
tcpm_debugfs_init(const struct tcpm_port * port)893 static void tcpm_debugfs_init(const struct tcpm_port *port) { }
tcpm_debugfs_exit(const struct tcpm_port * port)894 static void tcpm_debugfs_exit(const struct tcpm_port *port) { }
895
896 #endif
897
tcpm_set_cc(struct tcpm_port * port,enum typec_cc_status cc)898 static void tcpm_set_cc(struct tcpm_port *port, enum typec_cc_status cc)
899 {
900 tcpm_log(port, "cc:=%d", cc);
901 port->cc_req = cc;
902 port->tcpc->set_cc(port->tcpc, cc);
903 }
904
tcpm_enable_auto_vbus_discharge(struct tcpm_port * port,bool enable)905 static int tcpm_enable_auto_vbus_discharge(struct tcpm_port *port, bool enable)
906 {
907 int ret = 0;
908
909 if (port->tcpc->enable_auto_vbus_discharge) {
910 ret = port->tcpc->enable_auto_vbus_discharge(port->tcpc, enable);
911 tcpm_log_force(port, "%s vbus discharge ret:%d",
912 str_enable_disable(enable), ret);
913 if (!ret)
914 port->auto_vbus_discharge_enabled = enable;
915 }
916
917 return ret;
918 }
919
tcpm_apply_rc(struct tcpm_port * port)920 static void tcpm_apply_rc(struct tcpm_port *port)
921 {
922 /*
923 * TCPCI: Move to APPLY_RC state to prevent disconnect during PR_SWAP
924 * when Vbus auto discharge on disconnect is enabled.
925 */
926 if (port->tcpc->enable_auto_vbus_discharge && port->tcpc->apply_rc) {
927 tcpm_log(port, "Apply_RC");
928 port->tcpc->apply_rc(port->tcpc, port->cc_req, port->polarity);
929 tcpm_enable_auto_vbus_discharge(port, false);
930 }
931 }
932
933 /*
934 * Determine RP value to set based on maximum current supported
935 * by a port if configured as source.
936 * Returns CC value to report to link partner.
937 */
tcpm_rp_cc(struct tcpm_port * port)938 static enum typec_cc_status tcpm_rp_cc(struct tcpm_port *port)
939 {
940 const u32 *src_pdo = port->src_pdo;
941 int nr_pdo = port->nr_src_pdo;
942 int i;
943
944 if (!port->pd_supported)
945 return port->src_rp;
946
947 /*
948 * Search for first entry with matching voltage.
949 * It should report the maximum supported current.
950 */
951 for (i = 0; i < nr_pdo; i++) {
952 const u32 pdo = src_pdo[i];
953
954 if (pdo_type(pdo) == PDO_TYPE_FIXED &&
955 pdo_fixed_voltage(pdo) == 5000) {
956 unsigned int curr = pdo_max_current(pdo);
957
958 if (curr >= 3000)
959 return TYPEC_CC_RP_3_0;
960 else if (curr >= 1500)
961 return TYPEC_CC_RP_1_5;
962 return TYPEC_CC_RP_DEF;
963 }
964 }
965
966 return TYPEC_CC_RP_DEF;
967 }
968
tcpm_ams_finish(struct tcpm_port * port)969 static void tcpm_ams_finish(struct tcpm_port *port)
970 {
971 tcpm_log(port, "AMS %s finished", tcpm_ams_str[port->ams]);
972
973 if (port->pd_capable && port->pwr_role == TYPEC_SOURCE) {
974 if (port->negotiated_rev >= PD_REV30)
975 tcpm_set_cc(port, SINK_TX_OK);
976 else
977 tcpm_set_cc(port, SINK_TX_NG);
978 } else if (port->pwr_role == TYPEC_SOURCE) {
979 tcpm_set_cc(port, tcpm_rp_cc(port));
980 }
981
982 port->in_ams = false;
983 port->ams = NONE_AMS;
984 }
985
tcpm_pd_transmit(struct tcpm_port * port,enum tcpm_transmit_type tx_sop_type,const struct pd_message * msg)986 static int tcpm_pd_transmit(struct tcpm_port *port,
987 enum tcpm_transmit_type tx_sop_type,
988 const struct pd_message *msg)
989 {
990 unsigned long time_left;
991 int ret;
992 unsigned int negotiated_rev;
993
994 switch (tx_sop_type) {
995 case TCPC_TX_SOP_PRIME:
996 negotiated_rev = port->negotiated_rev_prime;
997 break;
998 case TCPC_TX_SOP:
999 default:
1000 negotiated_rev = port->negotiated_rev;
1001 break;
1002 }
1003
1004 if (msg)
1005 tcpm_log(port, "PD TX, header: %#x", le16_to_cpu(msg->header));
1006 else
1007 tcpm_log(port, "PD TX, type: %#x", tx_sop_type);
1008
1009 reinit_completion(&port->tx_complete);
1010 ret = port->tcpc->pd_transmit(port->tcpc, tx_sop_type, msg, negotiated_rev);
1011 if (ret < 0)
1012 return ret;
1013
1014 mutex_unlock(&port->lock);
1015 time_left = wait_for_completion_timeout(&port->tx_complete,
1016 msecs_to_jiffies(PD_T_TCPC_TX_TIMEOUT));
1017 mutex_lock(&port->lock);
1018 if (!time_left)
1019 return -ETIMEDOUT;
1020
1021 switch (port->tx_status) {
1022 case TCPC_TX_SUCCESS:
1023 switch (tx_sop_type) {
1024 case TCPC_TX_SOP_PRIME:
1025 port->message_id_prime = (port->message_id_prime + 1) &
1026 PD_HEADER_ID_MASK;
1027 break;
1028 case TCPC_TX_SOP:
1029 default:
1030 port->message_id = (port->message_id + 1) &
1031 PD_HEADER_ID_MASK;
1032 break;
1033 }
1034 /*
1035 * USB PD rev 2.0, 8.3.2.2.1:
1036 * USB PD rev 3.0, 8.3.2.1.3:
1037 * "... Note that every AMS is Interruptible until the first
1038 * Message in the sequence has been successfully sent (GoodCRC
1039 * Message received)."
1040 */
1041 if (port->ams != NONE_AMS)
1042 port->in_ams = true;
1043 break;
1044 case TCPC_TX_DISCARDED:
1045 ret = -EAGAIN;
1046 break;
1047 case TCPC_TX_FAILED:
1048 default:
1049 ret = -EIO;
1050 break;
1051 }
1052
1053 /* Some AMS don't expect responses. Finish them here. */
1054 if (port->ams == ATTENTION || port->ams == SOURCE_ALERT)
1055 tcpm_ams_finish(port);
1056
1057 return ret;
1058 }
1059
tcpm_pd_transmit_complete(struct tcpm_port * port,enum tcpm_transmit_status status)1060 void tcpm_pd_transmit_complete(struct tcpm_port *port,
1061 enum tcpm_transmit_status status)
1062 {
1063 tcpm_log(port, "PD TX complete, status: %u", status);
1064 port->tx_status = status;
1065 complete(&port->tx_complete);
1066 }
1067 EXPORT_SYMBOL_GPL(tcpm_pd_transmit_complete);
1068
tcpm_mux_set(struct tcpm_port * port,int state,enum usb_role usb_role,enum typec_orientation orientation)1069 static int tcpm_mux_set(struct tcpm_port *port, int state,
1070 enum usb_role usb_role,
1071 enum typec_orientation orientation)
1072 {
1073 int ret;
1074
1075 tcpm_log(port, "Requesting mux state %d, usb-role %d, orientation %d",
1076 state, usb_role, orientation);
1077
1078 ret = typec_set_orientation(port->typec_port, orientation);
1079 if (ret)
1080 return ret;
1081
1082 if (port->role_sw) {
1083 ret = usb_role_switch_set_role(port->role_sw, usb_role);
1084 if (ret)
1085 return ret;
1086 }
1087
1088 return typec_set_mode(port->typec_port, state);
1089 }
1090
tcpm_set_polarity(struct tcpm_port * port,enum typec_cc_polarity polarity)1091 static int tcpm_set_polarity(struct tcpm_port *port,
1092 enum typec_cc_polarity polarity)
1093 {
1094 int ret;
1095
1096 tcpm_log(port, "polarity %d", polarity);
1097
1098 ret = port->tcpc->set_polarity(port->tcpc, polarity);
1099 if (ret < 0)
1100 return ret;
1101
1102 port->polarity = polarity;
1103
1104 return 0;
1105 }
1106
tcpm_set_vconn(struct tcpm_port * port,bool enable)1107 static int tcpm_set_vconn(struct tcpm_port *port, bool enable)
1108 {
1109 int ret;
1110
1111 tcpm_log(port, "vconn:=%d", enable);
1112
1113 ret = port->tcpc->set_vconn(port->tcpc, enable);
1114 if (!ret) {
1115 port->vconn_role = enable ? TYPEC_SOURCE : TYPEC_SINK;
1116 typec_set_vconn_role(port->typec_port, port->vconn_role);
1117 }
1118
1119 return ret;
1120 }
1121
tcpm_get_current_limit(struct tcpm_port * port)1122 static u32 tcpm_get_current_limit(struct tcpm_port *port)
1123 {
1124 enum typec_cc_status cc;
1125 u32 limit;
1126
1127 cc = port->polarity ? port->cc2 : port->cc1;
1128 switch (cc) {
1129 case TYPEC_CC_RP_1_5:
1130 limit = 1500;
1131 break;
1132 case TYPEC_CC_RP_3_0:
1133 limit = 3000;
1134 break;
1135 case TYPEC_CC_RP_DEF:
1136 default:
1137 if (port->tcpc->get_current_limit)
1138 limit = port->tcpc->get_current_limit(port->tcpc);
1139 else
1140 limit = 0;
1141 break;
1142 }
1143
1144 return limit;
1145 }
1146
tcpm_set_current_limit(struct tcpm_port * port,u32 max_ma,u32 mv)1147 static int tcpm_set_current_limit(struct tcpm_port *port, u32 max_ma, u32 mv)
1148 {
1149 int ret = -EOPNOTSUPP;
1150
1151 tcpm_log(port, "Setting voltage/current limit %u mV %u mA", mv, max_ma);
1152
1153 port->supply_voltage = mv;
1154 port->current_limit = max_ma;
1155 power_supply_changed(port->psy);
1156
1157 if (port->tcpc->set_current_limit)
1158 ret = port->tcpc->set_current_limit(port->tcpc, max_ma, mv);
1159
1160 return ret;
1161 }
1162
tcpm_set_attached_state(struct tcpm_port * port,bool attached)1163 static int tcpm_set_attached_state(struct tcpm_port *port, bool attached)
1164 {
1165 return port->tcpc->set_roles(port->tcpc, attached, port->pwr_role,
1166 port->data_role);
1167 }
1168
tcpm_set_roles(struct tcpm_port * port,bool attached,int state,enum typec_role role,enum typec_data_role data)1169 static int tcpm_set_roles(struct tcpm_port *port, bool attached, int state,
1170 enum typec_role role, enum typec_data_role data)
1171 {
1172 enum typec_orientation orientation;
1173 enum usb_role usb_role;
1174 int ret;
1175
1176 if (port->polarity == TYPEC_POLARITY_CC1)
1177 orientation = TYPEC_ORIENTATION_NORMAL;
1178 else
1179 orientation = TYPEC_ORIENTATION_REVERSE;
1180
1181 if (port->typec_caps.data == TYPEC_PORT_DRD) {
1182 if (data == TYPEC_HOST)
1183 usb_role = USB_ROLE_HOST;
1184 else
1185 usb_role = USB_ROLE_DEVICE;
1186 } else if (port->typec_caps.data == TYPEC_PORT_DFP) {
1187 if (data == TYPEC_HOST) {
1188 if (role == TYPEC_SOURCE)
1189 usb_role = USB_ROLE_HOST;
1190 else
1191 usb_role = USB_ROLE_NONE;
1192 } else {
1193 return -ENOTSUPP;
1194 }
1195 } else {
1196 if (data == TYPEC_DEVICE) {
1197 if (role == TYPEC_SINK)
1198 usb_role = USB_ROLE_DEVICE;
1199 else
1200 usb_role = USB_ROLE_NONE;
1201 } else {
1202 return -ENOTSUPP;
1203 }
1204 }
1205
1206 ret = tcpm_mux_set(port, state, usb_role, orientation);
1207 if (ret < 0)
1208 return ret;
1209
1210 ret = port->tcpc->set_roles(port->tcpc, attached, role, data);
1211 if (ret < 0)
1212 return ret;
1213
1214 if (port->tcpc->set_orientation) {
1215 ret = port->tcpc->set_orientation(port->tcpc, orientation);
1216 if (ret < 0)
1217 return ret;
1218 }
1219
1220 port->pwr_role = role;
1221 port->data_role = data;
1222 typec_set_data_role(port->typec_port, data);
1223 typec_set_pwr_role(port->typec_port, role);
1224
1225 return 0;
1226 }
1227
tcpm_set_pwr_role(struct tcpm_port * port,enum typec_role role)1228 static int tcpm_set_pwr_role(struct tcpm_port *port, enum typec_role role)
1229 {
1230 int ret;
1231
1232 ret = port->tcpc->set_roles(port->tcpc, true, role,
1233 port->data_role);
1234 if (ret < 0)
1235 return ret;
1236
1237 port->pwr_role = role;
1238 typec_set_pwr_role(port->typec_port, role);
1239
1240 return 0;
1241 }
1242
1243 /*
1244 * Transform the PDO to be compliant to PD rev2.0.
1245 * Return 0 if the PDO type is not defined in PD rev2.0.
1246 * Otherwise, return the converted PDO.
1247 */
tcpm_forge_legacy_pdo(struct tcpm_port * port,u32 pdo,enum typec_role role)1248 static u32 tcpm_forge_legacy_pdo(struct tcpm_port *port, u32 pdo, enum typec_role role)
1249 {
1250 switch (pdo_type(pdo)) {
1251 case PDO_TYPE_FIXED:
1252 if (role == TYPEC_SINK)
1253 return pdo & ~PDO_FIXED_FRS_CURR_MASK;
1254 else
1255 return pdo & ~PDO_FIXED_UNCHUNK_EXT;
1256 case PDO_TYPE_VAR:
1257 case PDO_TYPE_BATT:
1258 return pdo;
1259 case PDO_TYPE_APDO:
1260 default:
1261 return 0;
1262 }
1263 }
1264
tcpm_pd_send_revision(struct tcpm_port * port)1265 static int tcpm_pd_send_revision(struct tcpm_port *port)
1266 {
1267 struct pd_message msg;
1268 u32 rmdo;
1269
1270 memset(&msg, 0, sizeof(msg));
1271 rmdo = RMDO(port->pd_rev.rev_major, port->pd_rev.rev_minor,
1272 port->pd_rev.ver_major, port->pd_rev.ver_minor);
1273 msg.payload[0] = cpu_to_le32(rmdo);
1274 msg.header = PD_HEADER_LE(PD_DATA_REVISION,
1275 port->pwr_role,
1276 port->data_role,
1277 port->negotiated_rev,
1278 port->message_id,
1279 1);
1280 return tcpm_pd_transmit(port, TCPC_TX_SOP, &msg);
1281 }
1282
tcpm_pd_send_source_caps(struct tcpm_port * port)1283 static int tcpm_pd_send_source_caps(struct tcpm_port *port)
1284 {
1285 struct pd_message msg;
1286 u32 pdo;
1287 unsigned int i, nr_pdo = 0;
1288
1289 memset(&msg, 0, sizeof(msg));
1290
1291 for (i = 0; i < port->nr_src_pdo; i++) {
1292 if (port->negotiated_rev >= PD_REV30) {
1293 msg.payload[nr_pdo++] = cpu_to_le32(port->src_pdo[i]);
1294 } else {
1295 pdo = tcpm_forge_legacy_pdo(port, port->src_pdo[i], TYPEC_SOURCE);
1296 if (pdo)
1297 msg.payload[nr_pdo++] = cpu_to_le32(pdo);
1298 }
1299 }
1300
1301 if (!nr_pdo) {
1302 /* No source capabilities defined, sink only */
1303 msg.header = PD_HEADER_LE(PD_CTRL_REJECT,
1304 port->pwr_role,
1305 port->data_role,
1306 port->negotiated_rev,
1307 port->message_id, 0);
1308 } else {
1309 msg.header = PD_HEADER_LE(PD_DATA_SOURCE_CAP,
1310 port->pwr_role,
1311 port->data_role,
1312 port->negotiated_rev,
1313 port->message_id,
1314 nr_pdo);
1315 }
1316
1317 return tcpm_pd_transmit(port, TCPC_TX_SOP, &msg);
1318 }
1319
tcpm_pd_send_sink_caps(struct tcpm_port * port)1320 static int tcpm_pd_send_sink_caps(struct tcpm_port *port)
1321 {
1322 struct pd_message msg;
1323 u32 pdo;
1324 unsigned int i, nr_pdo = 0;
1325
1326 memset(&msg, 0, sizeof(msg));
1327
1328 for (i = 0; i < port->nr_snk_pdo; i++) {
1329 if (port->negotiated_rev >= PD_REV30) {
1330 msg.payload[nr_pdo++] = cpu_to_le32(port->snk_pdo[i]);
1331 } else {
1332 pdo = tcpm_forge_legacy_pdo(port, port->snk_pdo[i], TYPEC_SINK);
1333 if (pdo)
1334 msg.payload[nr_pdo++] = cpu_to_le32(pdo);
1335 }
1336 }
1337
1338 if (!nr_pdo) {
1339 /* No sink capabilities defined, source only */
1340 msg.header = PD_HEADER_LE(PD_CTRL_REJECT,
1341 port->pwr_role,
1342 port->data_role,
1343 port->negotiated_rev,
1344 port->message_id, 0);
1345 } else {
1346 msg.header = PD_HEADER_LE(PD_DATA_SINK_CAP,
1347 port->pwr_role,
1348 port->data_role,
1349 port->negotiated_rev,
1350 port->message_id,
1351 nr_pdo);
1352 }
1353
1354 return tcpm_pd_transmit(port, TCPC_TX_SOP, &msg);
1355 }
1356
mod_tcpm_delayed_work(struct tcpm_port * port,unsigned int delay_ms)1357 static void mod_tcpm_delayed_work(struct tcpm_port *port, unsigned int delay_ms)
1358 {
1359 if (delay_ms) {
1360 hrtimer_start(&port->state_machine_timer, ms_to_ktime(delay_ms), HRTIMER_MODE_REL);
1361 } else {
1362 hrtimer_cancel(&port->state_machine_timer);
1363 kthread_queue_work(port->wq, &port->state_machine);
1364 }
1365 }
1366
mod_vdm_delayed_work(struct tcpm_port * port,unsigned int delay_ms)1367 static void mod_vdm_delayed_work(struct tcpm_port *port, unsigned int delay_ms)
1368 {
1369 if (delay_ms) {
1370 hrtimer_start(&port->vdm_state_machine_timer, ms_to_ktime(delay_ms),
1371 HRTIMER_MODE_REL);
1372 } else {
1373 hrtimer_cancel(&port->vdm_state_machine_timer);
1374 kthread_queue_work(port->wq, &port->vdm_state_machine);
1375 }
1376 }
1377
mod_enable_frs_delayed_work(struct tcpm_port * port,unsigned int delay_ms)1378 static void mod_enable_frs_delayed_work(struct tcpm_port *port, unsigned int delay_ms)
1379 {
1380 if (delay_ms) {
1381 hrtimer_start(&port->enable_frs_timer, ms_to_ktime(delay_ms), HRTIMER_MODE_REL);
1382 } else {
1383 hrtimer_cancel(&port->enable_frs_timer);
1384 kthread_queue_work(port->wq, &port->enable_frs);
1385 }
1386 }
1387
mod_send_discover_delayed_work(struct tcpm_port * port,unsigned int delay_ms)1388 static void mod_send_discover_delayed_work(struct tcpm_port *port, unsigned int delay_ms)
1389 {
1390 if (delay_ms) {
1391 hrtimer_start(&port->send_discover_timer, ms_to_ktime(delay_ms), HRTIMER_MODE_REL);
1392 } else {
1393 hrtimer_cancel(&port->send_discover_timer);
1394 kthread_queue_work(port->wq, &port->send_discover_work);
1395 }
1396 }
1397
tcpm_set_state(struct tcpm_port * port,enum tcpm_state state,unsigned int delay_ms)1398 static void tcpm_set_state(struct tcpm_port *port, enum tcpm_state state,
1399 unsigned int delay_ms)
1400 {
1401 if (delay_ms) {
1402 tcpm_log(port, "pending state change %s -> %s @ %u ms [%s %s]",
1403 tcpm_states[port->state], tcpm_states[state], delay_ms,
1404 pd_rev[port->negotiated_rev], tcpm_ams_str[port->ams]);
1405 port->delayed_state = state;
1406 mod_tcpm_delayed_work(port, delay_ms);
1407 port->delayed_runtime = ktime_add(ktime_get(), ms_to_ktime(delay_ms));
1408 port->delay_ms = delay_ms;
1409 } else {
1410 tcpm_log(port, "state change %s -> %s [%s %s]",
1411 tcpm_states[port->state], tcpm_states[state],
1412 pd_rev[port->negotiated_rev], tcpm_ams_str[port->ams]);
1413 port->delayed_state = INVALID_STATE;
1414 port->prev_state = port->state;
1415 port->state = state;
1416 /*
1417 * Don't re-queue the state machine work item if we're currently
1418 * in the state machine and we're immediately changing states.
1419 * tcpm_state_machine_work() will continue running the state
1420 * machine.
1421 */
1422 if (!port->state_machine_running)
1423 mod_tcpm_delayed_work(port, 0);
1424 }
1425 }
1426
tcpm_set_state_cond(struct tcpm_port * port,enum tcpm_state state,unsigned int delay_ms)1427 static void tcpm_set_state_cond(struct tcpm_port *port, enum tcpm_state state,
1428 unsigned int delay_ms)
1429 {
1430 if (port->enter_state == port->state)
1431 tcpm_set_state(port, state, delay_ms);
1432 else
1433 tcpm_log(port,
1434 "skipped %sstate change %s -> %s [%u ms], context state %s [%s %s]",
1435 delay_ms ? "delayed " : "",
1436 tcpm_states[port->state], tcpm_states[state],
1437 delay_ms, tcpm_states[port->enter_state],
1438 pd_rev[port->negotiated_rev], tcpm_ams_str[port->ams]);
1439 }
1440
tcpm_queue_message(struct tcpm_port * port,enum pd_msg_request message)1441 static void tcpm_queue_message(struct tcpm_port *port,
1442 enum pd_msg_request message)
1443 {
1444 port->queued_message = message;
1445 mod_tcpm_delayed_work(port, 0);
1446 }
1447
tcpm_vdm_ams(struct tcpm_port * port)1448 static bool tcpm_vdm_ams(struct tcpm_port *port)
1449 {
1450 switch (port->ams) {
1451 case DISCOVER_IDENTITY:
1452 case SOURCE_STARTUP_CABLE_PLUG_DISCOVER_IDENTITY:
1453 case DISCOVER_SVIDS:
1454 case DISCOVER_MODES:
1455 case DFP_TO_UFP_ENTER_MODE:
1456 case DFP_TO_UFP_EXIT_MODE:
1457 case DFP_TO_CABLE_PLUG_ENTER_MODE:
1458 case DFP_TO_CABLE_PLUG_EXIT_MODE:
1459 case ATTENTION:
1460 case UNSTRUCTURED_VDMS:
1461 case STRUCTURED_VDMS:
1462 break;
1463 default:
1464 return false;
1465 }
1466
1467 return true;
1468 }
1469
tcpm_ams_interruptible(struct tcpm_port * port)1470 static bool tcpm_ams_interruptible(struct tcpm_port *port)
1471 {
1472 switch (port->ams) {
1473 /* Interruptible AMS */
1474 case NONE_AMS:
1475 case SECURITY:
1476 case FIRMWARE_UPDATE:
1477 case DISCOVER_IDENTITY:
1478 case SOURCE_STARTUP_CABLE_PLUG_DISCOVER_IDENTITY:
1479 case DISCOVER_SVIDS:
1480 case DISCOVER_MODES:
1481 case DFP_TO_UFP_ENTER_MODE:
1482 case DFP_TO_UFP_EXIT_MODE:
1483 case DFP_TO_CABLE_PLUG_ENTER_MODE:
1484 case DFP_TO_CABLE_PLUG_EXIT_MODE:
1485 case UNSTRUCTURED_VDMS:
1486 case STRUCTURED_VDMS:
1487 case COUNTRY_INFO:
1488 case COUNTRY_CODES:
1489 break;
1490 /* Non-Interruptible AMS */
1491 default:
1492 if (port->in_ams)
1493 return false;
1494 break;
1495 }
1496
1497 return true;
1498 }
1499
tcpm_ams_start(struct tcpm_port * port,enum tcpm_ams ams)1500 static int tcpm_ams_start(struct tcpm_port *port, enum tcpm_ams ams)
1501 {
1502 int ret = 0;
1503
1504 tcpm_log(port, "AMS %s start", tcpm_ams_str[ams]);
1505
1506 if (!tcpm_ams_interruptible(port) &&
1507 !(ams == HARD_RESET || ams == SOFT_RESET_AMS)) {
1508 port->upcoming_state = INVALID_STATE;
1509 tcpm_log(port, "AMS %s not interruptible, aborting",
1510 tcpm_ams_str[port->ams]);
1511 return -EAGAIN;
1512 }
1513
1514 if (port->pwr_role == TYPEC_SOURCE) {
1515 enum typec_cc_status cc_req = port->cc_req;
1516
1517 port->ams = ams;
1518
1519 if (ams == HARD_RESET) {
1520 tcpm_set_cc(port, tcpm_rp_cc(port));
1521 tcpm_pd_transmit(port, TCPC_TX_HARD_RESET, NULL);
1522 tcpm_set_state(port, HARD_RESET_START, 0);
1523 return ret;
1524 } else if (ams == SOFT_RESET_AMS) {
1525 if (!port->explicit_contract)
1526 tcpm_set_cc(port, tcpm_rp_cc(port));
1527 tcpm_set_state(port, SOFT_RESET_SEND, 0);
1528 return ret;
1529 } else if (tcpm_vdm_ams(port)) {
1530 /* tSinkTx is enforced in vdm_run_state_machine */
1531 if (port->negotiated_rev >= PD_REV30)
1532 tcpm_set_cc(port, SINK_TX_NG);
1533 return ret;
1534 }
1535
1536 if (port->negotiated_rev >= PD_REV30)
1537 tcpm_set_cc(port, SINK_TX_NG);
1538
1539 switch (port->state) {
1540 case SRC_READY:
1541 case SRC_STARTUP:
1542 case SRC_SOFT_RESET_WAIT_SNK_TX:
1543 case SOFT_RESET:
1544 case SOFT_RESET_SEND:
1545 if (port->negotiated_rev >= PD_REV30)
1546 tcpm_set_state(port, AMS_START,
1547 cc_req == SINK_TX_OK ?
1548 PD_T_SINK_TX : 0);
1549 else
1550 tcpm_set_state(port, AMS_START, 0);
1551 break;
1552 default:
1553 if (port->negotiated_rev >= PD_REV30)
1554 tcpm_set_state(port, SRC_READY,
1555 cc_req == SINK_TX_OK ?
1556 PD_T_SINK_TX : 0);
1557 else
1558 tcpm_set_state(port, SRC_READY, 0);
1559 break;
1560 }
1561 } else {
1562 if (port->negotiated_rev >= PD_REV30 &&
1563 !tcpm_sink_tx_ok(port) &&
1564 ams != SOFT_RESET_AMS &&
1565 ams != HARD_RESET) {
1566 port->upcoming_state = INVALID_STATE;
1567 tcpm_log(port, "Sink TX No Go");
1568 return -EAGAIN;
1569 }
1570
1571 port->ams = ams;
1572
1573 if (ams == HARD_RESET) {
1574 tcpm_pd_transmit(port, TCPC_TX_HARD_RESET, NULL);
1575 tcpm_set_state(port, HARD_RESET_START, 0);
1576 return ret;
1577 } else if (tcpm_vdm_ams(port)) {
1578 return ret;
1579 }
1580
1581 if (port->state == SNK_READY ||
1582 port->state == SNK_SOFT_RESET)
1583 tcpm_set_state(port, AMS_START, 0);
1584 else
1585 tcpm_set_state(port, SNK_READY, 0);
1586 }
1587
1588 return ret;
1589 }
1590
1591 /*
1592 * VDM/VDO handling functions
1593 */
tcpm_queue_vdm(struct tcpm_port * port,const u32 header,const u32 * data,int cnt,enum tcpm_transmit_type tx_sop_type)1594 static void tcpm_queue_vdm(struct tcpm_port *port, const u32 header,
1595 const u32 *data, int cnt, enum tcpm_transmit_type tx_sop_type)
1596 {
1597 u32 vdo_hdr = port->vdo_data[0];
1598
1599 WARN_ON(!mutex_is_locked(&port->lock));
1600
1601 /* If is sending discover_identity, handle received message first */
1602 if (PD_VDO_SVDM(vdo_hdr) && PD_VDO_CMD(vdo_hdr) == CMD_DISCOVER_IDENT) {
1603 if (tx_sop_type == TCPC_TX_SOP_PRIME)
1604 port->send_discover_prime = true;
1605 else
1606 port->send_discover = true;
1607 mod_send_discover_delayed_work(port, SEND_DISCOVER_RETRY_MS);
1608 } else {
1609 /* Make sure we are not still processing a previous VDM packet */
1610 WARN_ON(port->vdm_state > VDM_STATE_DONE);
1611 }
1612
1613 port->vdo_count = cnt + 1;
1614 port->vdo_data[0] = header;
1615 memcpy(&port->vdo_data[1], data, sizeof(u32) * cnt);
1616 /* Set ready, vdm state machine will actually send */
1617 port->vdm_retries = 0;
1618 port->vdm_state = VDM_STATE_READY;
1619 port->vdm_sm_running = true;
1620
1621 port->tx_sop_type = tx_sop_type;
1622
1623 mod_vdm_delayed_work(port, 0);
1624 }
1625
tcpm_queue_vdm_work(struct kthread_work * work)1626 static void tcpm_queue_vdm_work(struct kthread_work *work)
1627 {
1628 struct altmode_vdm_event *event = container_of(work,
1629 struct altmode_vdm_event,
1630 work);
1631 struct tcpm_port *port = event->port;
1632
1633 mutex_lock(&port->lock);
1634 if (port->state != SRC_READY && port->state != SNK_READY &&
1635 port->state != SRC_VDM_IDENTITY_REQUEST) {
1636 tcpm_log_force(port, "dropping altmode_vdm_event");
1637 goto port_unlock;
1638 }
1639
1640 tcpm_queue_vdm(port, event->header, event->data, event->cnt, event->tx_sop_type);
1641
1642 port_unlock:
1643 kfree(event->data);
1644 kfree(event);
1645 mutex_unlock(&port->lock);
1646 }
1647
tcpm_queue_vdm_unlocked(struct tcpm_port * port,const u32 header,const u32 * data,int cnt,enum tcpm_transmit_type tx_sop_type)1648 static int tcpm_queue_vdm_unlocked(struct tcpm_port *port, const u32 header,
1649 const u32 *data, int cnt, enum tcpm_transmit_type tx_sop_type)
1650 {
1651 struct altmode_vdm_event *event;
1652 u32 *data_cpy;
1653 int ret = -ENOMEM;
1654
1655 event = kzalloc(sizeof(*event), GFP_KERNEL);
1656 if (!event)
1657 goto err_event;
1658
1659 data_cpy = kcalloc(cnt, sizeof(u32), GFP_KERNEL);
1660 if (!data_cpy)
1661 goto err_data;
1662
1663 kthread_init_work(&event->work, tcpm_queue_vdm_work);
1664 event->port = port;
1665 event->header = header;
1666 memcpy(data_cpy, data, sizeof(u32) * cnt);
1667 event->data = data_cpy;
1668 event->cnt = cnt;
1669 event->tx_sop_type = tx_sop_type;
1670
1671 ret = kthread_queue_work(port->wq, &event->work);
1672 if (!ret) {
1673 ret = -EBUSY;
1674 goto err_queue;
1675 }
1676
1677 return 0;
1678
1679 err_queue:
1680 kfree(data_cpy);
1681 err_data:
1682 kfree(event);
1683 err_event:
1684 tcpm_log_force(port, "failed to queue altmode vdm, err:%d", ret);
1685 return ret;
1686 }
1687
svdm_consume_identity(struct tcpm_port * port,const u32 * p,int cnt)1688 static void svdm_consume_identity(struct tcpm_port *port, const u32 *p, int cnt)
1689 {
1690 u32 vdo = p[VDO_INDEX_IDH];
1691 u32 product = p[VDO_INDEX_PRODUCT];
1692
1693 memset(&port->mode_data, 0, sizeof(port->mode_data));
1694
1695 port->partner_ident.id_header = vdo;
1696 port->partner_ident.cert_stat = p[VDO_INDEX_CSTAT];
1697 port->partner_ident.product = product;
1698
1699 if (port->partner)
1700 typec_partner_set_identity(port->partner);
1701
1702 tcpm_log(port, "Identity: %04x:%04x.%04x",
1703 PD_IDH_VID(vdo),
1704 PD_PRODUCT_PID(product), product & 0xffff);
1705 }
1706
svdm_consume_identity_sop_prime(struct tcpm_port * port,const u32 * p,int cnt)1707 static void svdm_consume_identity_sop_prime(struct tcpm_port *port, const u32 *p, int cnt)
1708 {
1709 u32 idh = p[VDO_INDEX_IDH];
1710 u32 product = p[VDO_INDEX_PRODUCT];
1711 int svdm_version;
1712
1713 /*
1714 * Attempt to consume identity only if cable currently is not set
1715 */
1716 if (!IS_ERR_OR_NULL(port->cable))
1717 goto register_plug;
1718
1719 /* Reset cable identity */
1720 memset(&port->cable_ident, 0, sizeof(port->cable_ident));
1721
1722 /* Fill out id header, cert, product, cable VDO 1 */
1723 port->cable_ident.id_header = idh;
1724 port->cable_ident.cert_stat = p[VDO_INDEX_CSTAT];
1725 port->cable_ident.product = product;
1726 port->cable_ident.vdo[0] = p[VDO_INDEX_CABLE_1];
1727
1728 /* Fill out cable desc, infer svdm_version from pd revision */
1729 port->cable_desc.type = (enum typec_plug_type) (VDO_TYPEC_CABLE_TYPE(p[VDO_INDEX_CABLE_1]) +
1730 USB_PLUG_TYPE_A);
1731 port->cable_desc.active = PD_IDH_PTYPE(idh) == IDH_PTYPE_ACABLE ? 1 : 0;
1732 /* Log PD Revision and additional cable VDO from negotiated revision */
1733 switch (port->negotiated_rev_prime) {
1734 case PD_REV30:
1735 port->cable_desc.pd_revision = 0x0300;
1736 if (port->cable_desc.active)
1737 port->cable_ident.vdo[1] = p[VDO_INDEX_CABLE_2];
1738 break;
1739 case PD_REV20:
1740 port->cable_desc.pd_revision = 0x0200;
1741 break;
1742 default:
1743 port->cable_desc.pd_revision = 0x0200;
1744 break;
1745 }
1746 port->cable_desc.identity = &port->cable_ident;
1747 /* Register Cable, set identity and svdm_version */
1748 port->cable = typec_register_cable(port->typec_port, &port->cable_desc);
1749 if (IS_ERR_OR_NULL(port->cable))
1750 return;
1751 typec_cable_set_identity(port->cable);
1752 /* Get SVDM version */
1753 svdm_version = PD_VDO_SVDM_VER(p[VDO_INDEX_HDR]);
1754 typec_cable_set_svdm_version(port->cable, svdm_version);
1755
1756 register_plug:
1757 if (IS_ERR_OR_NULL(port->plug_prime)) {
1758 port->plug_prime_desc.index = TYPEC_PLUG_SOP_P;
1759 port->plug_prime = typec_register_plug(port->cable,
1760 &port->plug_prime_desc);
1761 }
1762 }
1763
svdm_consume_svids(struct tcpm_port * port,const u32 * p,int cnt,enum tcpm_transmit_type rx_sop_type)1764 static bool svdm_consume_svids(struct tcpm_port *port, const u32 *p, int cnt,
1765 enum tcpm_transmit_type rx_sop_type)
1766 {
1767 struct pd_mode_data *pmdata = rx_sop_type == TCPC_TX_SOP_PRIME ?
1768 &port->mode_data_prime : &port->mode_data;
1769 int i;
1770
1771 for (i = 1; i < cnt; i++) {
1772 u16 svid;
1773
1774 svid = (p[i] >> 16) & 0xffff;
1775 if (!svid)
1776 return false;
1777
1778 if (pmdata->nsvids >= SVID_DISCOVERY_MAX)
1779 goto abort;
1780
1781 pmdata->svids[pmdata->nsvids++] = svid;
1782 tcpm_log(port, "SVID %d: 0x%x", pmdata->nsvids, svid);
1783
1784 svid = p[i] & 0xffff;
1785 if (!svid)
1786 return false;
1787
1788 if (pmdata->nsvids >= SVID_DISCOVERY_MAX)
1789 goto abort;
1790
1791 pmdata->svids[pmdata->nsvids++] = svid;
1792 tcpm_log(port, "SVID %d: 0x%x", pmdata->nsvids, svid);
1793 }
1794
1795 /*
1796 * PD3.0 Spec 6.4.4.3.2: The SVIDs are returned 2 per VDO (see Table
1797 * 6-43), and can be returned maximum 6 VDOs per response (see Figure
1798 * 6-19). If the Respondersupports 12 or more SVID then the Discover
1799 * SVIDs Command Shall be executed multiple times until a Discover
1800 * SVIDs VDO is returned ending either with a SVID value of 0x0000 in
1801 * the last part of the last VDO or with a VDO containing two SVIDs
1802 * with values of 0x0000.
1803 *
1804 * However, some odd dockers support SVIDs less than 12 but without
1805 * 0x0000 in the last VDO, so we need to break the Discover SVIDs
1806 * request and return false here.
1807 */
1808 return cnt == 7;
1809 abort:
1810 tcpm_log(port, "SVID_DISCOVERY_MAX(%d) too low!", SVID_DISCOVERY_MAX);
1811 return false;
1812 }
1813
svdm_consume_modes(struct tcpm_port * port,const u32 * p,int cnt,enum tcpm_transmit_type rx_sop_type)1814 static void svdm_consume_modes(struct tcpm_port *port, const u32 *p, int cnt,
1815 enum tcpm_transmit_type rx_sop_type)
1816 {
1817 struct pd_mode_data *pmdata = &port->mode_data;
1818 struct typec_altmode_desc *paltmode;
1819 int i;
1820
1821 switch (rx_sop_type) {
1822 case TCPC_TX_SOP_PRIME:
1823 pmdata = &port->mode_data_prime;
1824 if (pmdata->altmodes >= ARRAY_SIZE(port->plug_prime_altmode)) {
1825 /* Already logged in svdm_consume_svids() */
1826 return;
1827 }
1828 break;
1829 case TCPC_TX_SOP:
1830 pmdata = &port->mode_data;
1831 if (pmdata->altmodes >= ARRAY_SIZE(port->partner_altmode)) {
1832 /* Already logged in svdm_consume_svids() */
1833 return;
1834 }
1835 break;
1836 default:
1837 return;
1838 }
1839
1840 for (i = 1; i < cnt; i++) {
1841 paltmode = &pmdata->altmode_desc[pmdata->altmodes];
1842 memset(paltmode, 0, sizeof(*paltmode));
1843
1844 paltmode->svid = pmdata->svids[pmdata->svid_index];
1845 paltmode->mode = i;
1846 paltmode->vdo = p[i];
1847
1848 tcpm_log(port, " Alternate mode %d: SVID 0x%04x, VDO %d: 0x%08x",
1849 pmdata->altmodes, paltmode->svid,
1850 paltmode->mode, paltmode->vdo);
1851
1852 pmdata->altmodes++;
1853 }
1854 }
1855
tcpm_register_partner_altmodes(struct tcpm_port * port)1856 static void tcpm_register_partner_altmodes(struct tcpm_port *port)
1857 {
1858 struct pd_mode_data *modep = &port->mode_data;
1859 struct typec_altmode *altmode;
1860 int i;
1861
1862 if (!port->partner)
1863 return;
1864
1865 for (i = 0; i < modep->altmodes; i++) {
1866 altmode = typec_partner_register_altmode(port->partner,
1867 &modep->altmode_desc[i]);
1868 if (IS_ERR(altmode)) {
1869 tcpm_log(port, "Failed to register partner SVID 0x%04x",
1870 modep->altmode_desc[i].svid);
1871 altmode = NULL;
1872 }
1873 port->partner_altmode[i] = altmode;
1874 }
1875 }
1876
tcpm_register_plug_altmodes(struct tcpm_port * port)1877 static void tcpm_register_plug_altmodes(struct tcpm_port *port)
1878 {
1879 struct pd_mode_data *modep = &port->mode_data_prime;
1880 struct typec_altmode *altmode;
1881 int i;
1882
1883 typec_plug_set_num_altmodes(port->plug_prime, modep->altmodes);
1884
1885 for (i = 0; i < modep->altmodes; i++) {
1886 altmode = typec_plug_register_altmode(port->plug_prime,
1887 &modep->altmode_desc[i]);
1888 if (IS_ERR(altmode)) {
1889 tcpm_log(port, "Failed to register plug SVID 0x%04x",
1890 modep->altmode_desc[i].svid);
1891 altmode = NULL;
1892 }
1893 port->plug_prime_altmode[i] = altmode;
1894 }
1895 }
1896
1897 #define supports_modal(port) PD_IDH_MODAL_SUPP((port)->partner_ident.id_header)
1898 #define supports_modal_cable(port) PD_IDH_MODAL_SUPP((port)->cable_ident.id_header)
1899 #define supports_host(port) PD_IDH_HOST_SUPP((port->partner_ident.id_header))
1900
1901 /*
1902 * Helper to determine whether the port is capable of SOP' communication at the
1903 * current point in time.
1904 */
tcpm_can_communicate_sop_prime(struct tcpm_port * port)1905 static bool tcpm_can_communicate_sop_prime(struct tcpm_port *port)
1906 {
1907 /* Check to see if tcpc supports SOP' communication */
1908 if (!port->tcpc->cable_comm_capable || !port->tcpc->cable_comm_capable(port->tcpc))
1909 return false;
1910 /*
1911 * Power Delivery 2.0 Section 6.3.11
1912 * Before communicating with a Cable Plug a Port Should ensure that it
1913 * is the Vconn Source and that the Cable Plugs are powered by
1914 * performing a Vconn swap if necessary. Since it cannot be guaranteed
1915 * that the present Vconn Source is supplying Vconn, the only means to
1916 * ensure that the Cable Plugs are powered is for a Port wishing to
1917 * communicate with a Cable Plug is to become the Vconn Source.
1918 *
1919 * Power Delivery 3.0 Section 6.3.11
1920 * Before communicating with a Cable Plug a Port Shall ensure that it
1921 * is the Vconn source.
1922 */
1923 if (port->vconn_role != TYPEC_SOURCE)
1924 return false;
1925 /*
1926 * Power Delivery 2.0 Section 2.4.4
1927 * When no Contract or an Implicit Contract is in place the Source can
1928 * communicate with a Cable Plug using SOP' packets in order to discover
1929 * its characteristics.
1930 *
1931 * Power Delivery 3.0 Section 2.4.4
1932 * When no Contract or an Implicit Contract is in place only the Source
1933 * port that is supplying Vconn is allowed to send packets to a Cable
1934 * Plug and is allowed to respond to packets from the Cable Plug.
1935 */
1936 if (!port->explicit_contract)
1937 return port->pwr_role == TYPEC_SOURCE;
1938 if (port->negotiated_rev == PD_REV30)
1939 return true;
1940 /*
1941 * Power Delivery 2.0 Section 2.4.4
1942 *
1943 * When an Explicit Contract is in place the DFP (either the Source or
1944 * the Sink) can communicate with the Cable Plug(s) using SOP’/SOP”
1945 * Packets (see Figure 2-3).
1946 */
1947 if (port->negotiated_rev == PD_REV20)
1948 return port->data_role == TYPEC_HOST;
1949 return false;
1950 }
1951
tcpm_attempt_vconn_swap_discovery(struct tcpm_port * port)1952 static bool tcpm_attempt_vconn_swap_discovery(struct tcpm_port *port)
1953 {
1954 if (!port->tcpc->attempt_vconn_swap_discovery)
1955 return false;
1956
1957 /* Port is already source, no need to perform swap */
1958 if (port->vconn_role == TYPEC_SOURCE)
1959 return false;
1960
1961 /*
1962 * Partner needs to support Alternate Modes with modal support. If
1963 * partner is also capable of being a USB Host, it could be a device
1964 * that supports Alternate Modes as the DFP.
1965 */
1966 if (!supports_modal(port) || supports_host(port))
1967 return false;
1968
1969 if ((port->negotiated_rev == PD_REV20 && port->data_role == TYPEC_HOST) ||
1970 port->negotiated_rev == PD_REV30)
1971 return port->tcpc->attempt_vconn_swap_discovery(port->tcpc);
1972
1973 return false;
1974 }
1975
1976
tcpm_cable_vdm_supported(struct tcpm_port * port)1977 static bool tcpm_cable_vdm_supported(struct tcpm_port *port)
1978 {
1979 return !IS_ERR_OR_NULL(port->cable) &&
1980 typec_cable_is_active(port->cable) &&
1981 supports_modal_cable(port) &&
1982 tcpm_can_communicate_sop_prime(port);
1983 }
1984
tcpm_pd_svdm(struct tcpm_port * port,struct typec_altmode * adev,const u32 * p,int cnt,u32 * response,enum adev_actions * adev_action,enum tcpm_transmit_type rx_sop_type,enum tcpm_transmit_type * response_tx_sop_type)1985 static int tcpm_pd_svdm(struct tcpm_port *port, struct typec_altmode *adev,
1986 const u32 *p, int cnt, u32 *response,
1987 enum adev_actions *adev_action,
1988 enum tcpm_transmit_type rx_sop_type,
1989 enum tcpm_transmit_type *response_tx_sop_type)
1990 {
1991 struct typec_port *typec = port->typec_port;
1992 struct typec_altmode *pdev, *pdev_prime;
1993 struct pd_mode_data *modep, *modep_prime;
1994 int svdm_version;
1995 int rlen = 0;
1996 int cmd_type;
1997 int cmd;
1998 int i;
1999 int ret;
2000
2001 cmd_type = PD_VDO_CMDT(p[0]);
2002 cmd = PD_VDO_CMD(p[0]);
2003
2004 tcpm_log(port, "Rx VDM cmd 0x%x type %d cmd %d len %d",
2005 p[0], cmd_type, cmd, cnt);
2006
2007 switch (rx_sop_type) {
2008 case TCPC_TX_SOP_PRIME:
2009 modep_prime = &port->mode_data_prime;
2010 pdev_prime = typec_match_altmode(port->plug_prime_altmode,
2011 ALTMODE_DISCOVERY_MAX,
2012 PD_VDO_VID(p[0]),
2013 PD_VDO_OPOS(p[0]));
2014 svdm_version = typec_get_cable_svdm_version(typec);
2015 /*
2016 * Update SVDM version if cable was discovered before port partner.
2017 */
2018 if (!IS_ERR_OR_NULL(port->cable) &&
2019 PD_VDO_SVDM_VER(p[0]) < svdm_version)
2020 typec_cable_set_svdm_version(port->cable, svdm_version);
2021 break;
2022 case TCPC_TX_SOP:
2023 modep = &port->mode_data;
2024 pdev = typec_match_altmode(port->partner_altmode,
2025 ALTMODE_DISCOVERY_MAX,
2026 PD_VDO_VID(p[0]),
2027 PD_VDO_OPOS(p[0]));
2028 svdm_version = typec_get_negotiated_svdm_version(typec);
2029 if (svdm_version < 0)
2030 return 0;
2031 break;
2032 default:
2033 modep = &port->mode_data;
2034 pdev = typec_match_altmode(port->partner_altmode,
2035 ALTMODE_DISCOVERY_MAX,
2036 PD_VDO_VID(p[0]),
2037 PD_VDO_OPOS(p[0]));
2038 svdm_version = typec_get_negotiated_svdm_version(typec);
2039 if (svdm_version < 0)
2040 return 0;
2041 break;
2042 }
2043
2044 switch (cmd_type) {
2045 case CMDT_INIT:
2046 /*
2047 * Only the port or port partner is allowed to initialize SVDM
2048 * commands over SOP'. In case the port partner initializes a
2049 * sequence when it is not allowed to send SOP' messages, drop
2050 * the message should the TCPM port try to process it.
2051 */
2052 if (rx_sop_type == TCPC_TX_SOP_PRIME)
2053 return 0;
2054
2055 switch (cmd) {
2056 case CMD_DISCOVER_IDENT:
2057 if (PD_VDO_VID(p[0]) != USB_SID_PD)
2058 break;
2059
2060 if (IS_ERR_OR_NULL(port->partner))
2061 break;
2062
2063 if (PD_VDO_SVDM_VER(p[0]) < svdm_version) {
2064 typec_partner_set_svdm_version(port->partner,
2065 PD_VDO_SVDM_VER(p[0]));
2066 svdm_version = PD_VDO_SVDM_VER(p[0]);
2067 }
2068
2069 port->ams = DISCOVER_IDENTITY;
2070 /*
2071 * PD2.0 Spec 6.10.3: respond with NAK as DFP (data host)
2072 * PD3.1 Spec 6.4.4.2.5.1: respond with NAK if "invalid field" or
2073 * "wrong configuation" or "Unrecognized"
2074 */
2075 if ((port->data_role == TYPEC_DEVICE || svdm_version >= SVDM_VER_2_0) &&
2076 port->nr_snk_vdo) {
2077 if (svdm_version < SVDM_VER_2_0) {
2078 for (i = 0; i < port->nr_snk_vdo_v1; i++)
2079 response[i + 1] = port->snk_vdo_v1[i];
2080 rlen = port->nr_snk_vdo_v1 + 1;
2081
2082 } else {
2083 for (i = 0; i < port->nr_snk_vdo; i++)
2084 response[i + 1] = port->snk_vdo[i];
2085 rlen = port->nr_snk_vdo + 1;
2086 }
2087 }
2088 break;
2089 case CMD_DISCOVER_SVID:
2090 port->ams = DISCOVER_SVIDS;
2091 break;
2092 case CMD_DISCOVER_MODES:
2093 port->ams = DISCOVER_MODES;
2094 break;
2095 case CMD_ENTER_MODE:
2096 port->ams = DFP_TO_UFP_ENTER_MODE;
2097 break;
2098 case CMD_EXIT_MODE:
2099 port->ams = DFP_TO_UFP_EXIT_MODE;
2100 break;
2101 case CMD_ATTENTION:
2102 /* Attention command does not have response */
2103 *adev_action = ADEV_ATTENTION;
2104 return 0;
2105 default:
2106 break;
2107 }
2108 if (rlen >= 1) {
2109 response[0] = p[0] | VDO_CMDT(CMDT_RSP_ACK);
2110 } else if (rlen == 0) {
2111 response[0] = p[0] | VDO_CMDT(CMDT_RSP_NAK);
2112 rlen = 1;
2113 } else {
2114 response[0] = p[0] | VDO_CMDT(CMDT_RSP_BUSY);
2115 rlen = 1;
2116 }
2117 response[0] = (response[0] & ~VDO_SVDM_VERS_MASK) |
2118 (VDO_SVDM_VERS(typec_get_negotiated_svdm_version(typec)));
2119 break;
2120 case CMDT_RSP_ACK:
2121 /*
2122 * Silently drop message if we are not connected, but can process
2123 * if SOP' Discover Identity prior to explicit contract.
2124 */
2125 if (IS_ERR_OR_NULL(port->partner) &&
2126 !(rx_sop_type == TCPC_TX_SOP_PRIME && cmd == CMD_DISCOVER_IDENT))
2127 break;
2128
2129 tcpm_ams_finish(port);
2130
2131 switch (cmd) {
2132 /*
2133 * SVDM Command Flow for SOP and SOP':
2134 * SOP Discover Identity
2135 * SOP' Discover Identity
2136 * SOP Discover SVIDs
2137 * Discover Modes
2138 * (Active Cables)
2139 * SOP' Discover SVIDs
2140 * Discover Modes
2141 *
2142 * Perform Discover SOP' if the port can communicate with cable
2143 * plug.
2144 */
2145 case CMD_DISCOVER_IDENT:
2146 switch (rx_sop_type) {
2147 case TCPC_TX_SOP:
2148 if (PD_VDO_SVDM_VER(p[0]) < svdm_version) {
2149 typec_partner_set_svdm_version(port->partner,
2150 PD_VDO_SVDM_VER(p[0]));
2151 /* If cable is discovered before partner, downgrade svdm */
2152 if (!IS_ERR_OR_NULL(port->cable) &&
2153 (typec_get_cable_svdm_version(port->typec_port) >
2154 svdm_version))
2155 typec_cable_set_svdm_version(port->cable,
2156 svdm_version);
2157 }
2158 /* 6.4.4.3.1 */
2159 svdm_consume_identity(port, p, cnt);
2160 /* Attempt Vconn swap, delay SOP' discovery if necessary */
2161 if (tcpm_attempt_vconn_swap_discovery(port)) {
2162 port->send_discover_prime = true;
2163 port->upcoming_state = VCONN_SWAP_SEND;
2164 ret = tcpm_ams_start(port, VCONN_SWAP);
2165 if (!ret)
2166 return 0;
2167 /* Cannot perform Vconn swap */
2168 port->upcoming_state = INVALID_STATE;
2169 port->send_discover_prime = false;
2170 }
2171
2172 /*
2173 * Attempt Discover Identity on SOP' if the
2174 * cable was not discovered previously, and use
2175 * the SVDM version of the partner to probe.
2176 */
2177 if (IS_ERR_OR_NULL(port->cable) &&
2178 tcpm_can_communicate_sop_prime(port)) {
2179 *response_tx_sop_type = TCPC_TX_SOP_PRIME;
2180 port->send_discover_prime = true;
2181 response[0] = VDO(USB_SID_PD, 1,
2182 typec_get_negotiated_svdm_version(typec),
2183 CMD_DISCOVER_IDENT);
2184 rlen = 1;
2185 } else {
2186 *response_tx_sop_type = TCPC_TX_SOP;
2187 response[0] = VDO(USB_SID_PD, 1,
2188 typec_get_negotiated_svdm_version(typec),
2189 CMD_DISCOVER_SVID);
2190 rlen = 1;
2191 }
2192 break;
2193 case TCPC_TX_SOP_PRIME:
2194 /*
2195 * svdm_consume_identity_sop_prime will determine
2196 * the svdm_version for the cable moving forward.
2197 */
2198 svdm_consume_identity_sop_prime(port, p, cnt);
2199
2200 /*
2201 * If received in SRC_VDM_IDENTITY_REQUEST, continue
2202 * to SRC_SEND_CAPABILITIES
2203 */
2204 if (port->state == SRC_VDM_IDENTITY_REQUEST) {
2205 tcpm_set_state(port, SRC_SEND_CAPABILITIES, 0);
2206 return 0;
2207 }
2208
2209 *response_tx_sop_type = TCPC_TX_SOP;
2210 response[0] = VDO(USB_SID_PD, 1,
2211 typec_get_negotiated_svdm_version(typec),
2212 CMD_DISCOVER_SVID);
2213 rlen = 1;
2214 break;
2215 default:
2216 return 0;
2217 }
2218 break;
2219 case CMD_DISCOVER_SVID:
2220 *response_tx_sop_type = rx_sop_type;
2221 /* 6.4.4.3.2 */
2222 if (svdm_consume_svids(port, p, cnt, rx_sop_type)) {
2223 response[0] = VDO(USB_SID_PD, 1, svdm_version, CMD_DISCOVER_SVID);
2224 rlen = 1;
2225 } else {
2226 if (rx_sop_type == TCPC_TX_SOP) {
2227 if (modep->nsvids && supports_modal(port)) {
2228 response[0] = VDO(modep->svids[0], 1, svdm_version,
2229 CMD_DISCOVER_MODES);
2230 rlen = 1;
2231 }
2232 } else if (rx_sop_type == TCPC_TX_SOP_PRIME) {
2233 if (modep_prime->nsvids) {
2234 response[0] = VDO(modep_prime->svids[0], 1,
2235 svdm_version, CMD_DISCOVER_MODES);
2236 rlen = 1;
2237 }
2238 }
2239 }
2240 break;
2241 case CMD_DISCOVER_MODES:
2242 if (rx_sop_type == TCPC_TX_SOP) {
2243 /* 6.4.4.3.3 */
2244 svdm_consume_modes(port, p, cnt, rx_sop_type);
2245 modep->svid_index++;
2246 if (modep->svid_index < modep->nsvids) {
2247 u16 svid = modep->svids[modep->svid_index];
2248 *response_tx_sop_type = TCPC_TX_SOP;
2249 response[0] = VDO(svid, 1, svdm_version,
2250 CMD_DISCOVER_MODES);
2251 rlen = 1;
2252 } else if (tcpm_cable_vdm_supported(port)) {
2253 *response_tx_sop_type = TCPC_TX_SOP_PRIME;
2254 response[0] = VDO(USB_SID_PD, 1,
2255 typec_get_cable_svdm_version(typec),
2256 CMD_DISCOVER_SVID);
2257 rlen = 1;
2258 } else {
2259 tcpm_register_partner_altmodes(port);
2260 }
2261 } else if (rx_sop_type == TCPC_TX_SOP_PRIME) {
2262 /* 6.4.4.3.3 */
2263 svdm_consume_modes(port, p, cnt, rx_sop_type);
2264 modep_prime->svid_index++;
2265 if (modep_prime->svid_index < modep_prime->nsvids) {
2266 u16 svid = modep_prime->svids[modep_prime->svid_index];
2267 *response_tx_sop_type = TCPC_TX_SOP_PRIME;
2268 response[0] = VDO(svid, 1,
2269 typec_get_cable_svdm_version(typec),
2270 CMD_DISCOVER_MODES);
2271 rlen = 1;
2272 } else {
2273 tcpm_register_plug_altmodes(port);
2274 tcpm_register_partner_altmodes(port);
2275 }
2276 }
2277 break;
2278 case CMD_ENTER_MODE:
2279 *response_tx_sop_type = rx_sop_type;
2280 if (rx_sop_type == TCPC_TX_SOP) {
2281 if (adev && pdev) {
2282 typec_altmode_update_active(pdev, true);
2283 *adev_action = ADEV_QUEUE_VDM_SEND_EXIT_MODE_ON_FAIL;
2284 }
2285 } else if (rx_sop_type == TCPC_TX_SOP_PRIME) {
2286 if (adev && pdev_prime) {
2287 typec_altmode_update_active(pdev_prime, true);
2288 *adev_action = ADEV_QUEUE_VDM_SEND_EXIT_MODE_ON_FAIL;
2289 }
2290 }
2291 return 0;
2292 case CMD_EXIT_MODE:
2293 *response_tx_sop_type = rx_sop_type;
2294 if (rx_sop_type == TCPC_TX_SOP) {
2295 if (adev && pdev) {
2296 typec_altmode_update_active(pdev, false);
2297 /* Back to USB Operation */
2298 *adev_action = ADEV_NOTIFY_USB_AND_QUEUE_VDM;
2299 return 0;
2300 }
2301 }
2302 break;
2303 case VDO_CMD_VENDOR(0) ... VDO_CMD_VENDOR(15):
2304 break;
2305 default:
2306 /* Unrecognized SVDM */
2307 response[0] = p[0] | VDO_CMDT(CMDT_RSP_NAK);
2308 rlen = 1;
2309 response[0] = (response[0] & ~VDO_SVDM_VERS_MASK) |
2310 (VDO_SVDM_VERS(svdm_version));
2311 break;
2312 }
2313 break;
2314 case CMDT_RSP_NAK:
2315 tcpm_ams_finish(port);
2316 switch (cmd) {
2317 case CMD_DISCOVER_IDENT:
2318 case CMD_DISCOVER_SVID:
2319 case CMD_DISCOVER_MODES:
2320 case VDO_CMD_VENDOR(0) ... VDO_CMD_VENDOR(15):
2321 break;
2322 case CMD_ENTER_MODE:
2323 /* Back to USB Operation */
2324 *adev_action = ADEV_NOTIFY_USB_AND_QUEUE_VDM;
2325 return 0;
2326 default:
2327 /* Unrecognized SVDM */
2328 response[0] = p[0] | VDO_CMDT(CMDT_RSP_NAK);
2329 rlen = 1;
2330 response[0] = (response[0] & ~VDO_SVDM_VERS_MASK) |
2331 (VDO_SVDM_VERS(svdm_version));
2332 break;
2333 }
2334 break;
2335 default:
2336 response[0] = p[0] | VDO_CMDT(CMDT_RSP_NAK);
2337 rlen = 1;
2338 response[0] = (response[0] & ~VDO_SVDM_VERS_MASK) |
2339 (VDO_SVDM_VERS(svdm_version));
2340 break;
2341 }
2342
2343 /* Informing the alternate mode drivers about everything */
2344 *adev_action = ADEV_QUEUE_VDM;
2345 return rlen;
2346 }
2347
2348 static void tcpm_pd_handle_msg(struct tcpm_port *port,
2349 enum pd_msg_request message,
2350 enum tcpm_ams ams);
2351
tcpm_handle_vdm_request(struct tcpm_port * port,const __le32 * payload,int cnt,enum tcpm_transmit_type rx_sop_type)2352 static void tcpm_handle_vdm_request(struct tcpm_port *port,
2353 const __le32 *payload, int cnt,
2354 enum tcpm_transmit_type rx_sop_type)
2355 {
2356 enum adev_actions adev_action = ADEV_NONE;
2357 struct typec_altmode *adev;
2358 u32 p[PD_MAX_PAYLOAD];
2359 u32 response[8] = { };
2360 int i, rlen = 0;
2361 enum tcpm_transmit_type response_tx_sop_type = TCPC_TX_SOP;
2362
2363 for (i = 0; i < cnt; i++)
2364 p[i] = le32_to_cpu(payload[i]);
2365
2366 adev = typec_match_altmode(port->port_altmode, ALTMODE_DISCOVERY_MAX,
2367 PD_VDO_VID(p[0]), PD_VDO_OPOS(p[0]));
2368
2369 if (port->vdm_state == VDM_STATE_BUSY) {
2370 /* If UFP responded busy retry after timeout */
2371 if (PD_VDO_CMDT(p[0]) == CMDT_RSP_BUSY) {
2372 port->vdm_state = VDM_STATE_WAIT_RSP_BUSY;
2373 port->vdo_retry = (p[0] & ~VDO_CMDT_MASK) |
2374 CMDT_INIT;
2375 mod_vdm_delayed_work(port, PD_T_VDM_BUSY);
2376 return;
2377 }
2378 port->vdm_state = VDM_STATE_DONE;
2379 }
2380
2381 if (PD_VDO_SVDM(p[0]) && (adev || tcpm_vdm_ams(port) || port->nr_snk_vdo)) {
2382 /*
2383 * Here a SVDM is received (INIT or RSP or unknown). Set the vdm_sm_running in
2384 * advance because we are dropping the lock but may send VDMs soon.
2385 * For the cases of INIT received:
2386 * - If no response to send, it will be cleared later in this function.
2387 * - If there are responses to send, it will be cleared in the state machine.
2388 * For the cases of RSP received:
2389 * - If no further INIT to send, it will be cleared later in this function.
2390 * - Otherwise, it will be cleared in the state machine if timeout or it will go
2391 * back here until no further INIT to send.
2392 * For the cases of unknown type received:
2393 * - We will send NAK and the flag will be cleared in the state machine.
2394 */
2395 port->vdm_sm_running = true;
2396 rlen = tcpm_pd_svdm(port, adev, p, cnt, response, &adev_action,
2397 rx_sop_type, &response_tx_sop_type);
2398 } else {
2399 if (port->negotiated_rev >= PD_REV30)
2400 tcpm_pd_handle_msg(port, PD_MSG_CTRL_NOT_SUPP, NONE_AMS);
2401 }
2402
2403 /*
2404 * We are done with any state stored in the port struct now, except
2405 * for any port struct changes done by the tcpm_queue_vdm() call
2406 * below, which is a separate operation.
2407 *
2408 * So we can safely release the lock here; and we MUST release the
2409 * lock here to avoid an AB BA lock inversion:
2410 *
2411 * If we keep the lock here then the lock ordering in this path is:
2412 * 1. tcpm_pd_rx_handler take the tcpm port lock
2413 * 2. One of the typec_altmode_* calls below takes the alt-mode's lock
2414 *
2415 * And we also have this ordering:
2416 * 1. alt-mode driver takes the alt-mode's lock
2417 * 2. alt-mode driver calls tcpm_altmode_enter which takes the
2418 * tcpm port lock
2419 *
2420 * Dropping our lock here avoids this.
2421 */
2422 mutex_unlock(&port->lock);
2423
2424 if (adev) {
2425 switch (adev_action) {
2426 case ADEV_NONE:
2427 break;
2428 case ADEV_NOTIFY_USB_AND_QUEUE_VDM:
2429 if (rx_sop_type == TCPC_TX_SOP_PRIME) {
2430 typec_cable_altmode_vdm(adev, TYPEC_PLUG_SOP_P, p[0], &p[1], cnt);
2431 } else {
2432 WARN_ON(typec_altmode_notify(adev, TYPEC_STATE_USB, NULL));
2433 typec_altmode_vdm(adev, p[0], &p[1], cnt);
2434 }
2435 break;
2436 case ADEV_QUEUE_VDM:
2437 if (rx_sop_type == TCPC_TX_SOP_PRIME)
2438 typec_cable_altmode_vdm(adev, TYPEC_PLUG_SOP_P, p[0], &p[1], cnt);
2439 else
2440 typec_altmode_vdm(adev, p[0], &p[1], cnt);
2441 break;
2442 case ADEV_QUEUE_VDM_SEND_EXIT_MODE_ON_FAIL:
2443 if (rx_sop_type == TCPC_TX_SOP_PRIME) {
2444 if (typec_cable_altmode_vdm(adev, TYPEC_PLUG_SOP_P,
2445 p[0], &p[1], cnt)) {
2446 int svdm_version = typec_get_cable_svdm_version(
2447 port->typec_port);
2448 if (svdm_version < 0)
2449 break;
2450
2451 response[0] = VDO(adev->svid, 1, svdm_version,
2452 CMD_EXIT_MODE);
2453 response[0] |= VDO_OPOS(adev->mode);
2454 rlen = 1;
2455 }
2456 } else {
2457 if (typec_altmode_vdm(adev, p[0], &p[1], cnt)) {
2458 int svdm_version = typec_get_negotiated_svdm_version(
2459 port->typec_port);
2460 if (svdm_version < 0)
2461 break;
2462
2463 response[0] = VDO(adev->svid, 1, svdm_version,
2464 CMD_EXIT_MODE);
2465 response[0] |= VDO_OPOS(adev->mode);
2466 rlen = 1;
2467 }
2468 }
2469 break;
2470 case ADEV_ATTENTION:
2471 if (typec_altmode_attention(adev, p[1]))
2472 tcpm_log(port, "typec_altmode_attention no port partner altmode");
2473 break;
2474 }
2475 }
2476
2477 /*
2478 * We must re-take the lock here to balance the unlock in
2479 * tcpm_pd_rx_handler, note that no changes, other then the
2480 * tcpm_queue_vdm call, are made while the lock is held again.
2481 * All that is done after the call is unwinding the call stack until
2482 * we return to tcpm_pd_rx_handler and do the unlock there.
2483 */
2484 mutex_lock(&port->lock);
2485
2486 if (rlen > 0)
2487 tcpm_queue_vdm(port, response[0], &response[1], rlen - 1, response_tx_sop_type);
2488 else
2489 port->vdm_sm_running = false;
2490 }
2491
tcpm_send_vdm(struct tcpm_port * port,u32 vid,int cmd,const u32 * data,int count,enum tcpm_transmit_type tx_sop_type)2492 static void tcpm_send_vdm(struct tcpm_port *port, u32 vid, int cmd,
2493 const u32 *data, int count, enum tcpm_transmit_type tx_sop_type)
2494 {
2495 int svdm_version;
2496 u32 header;
2497
2498 switch (tx_sop_type) {
2499 case TCPC_TX_SOP_PRIME:
2500 /*
2501 * If the port partner is discovered, then the port partner's
2502 * SVDM Version will be returned
2503 */
2504 svdm_version = typec_get_cable_svdm_version(port->typec_port);
2505 if (svdm_version < 0)
2506 svdm_version = SVDM_VER_MAX;
2507 break;
2508 case TCPC_TX_SOP:
2509 svdm_version = typec_get_negotiated_svdm_version(port->typec_port);
2510 if (svdm_version < 0)
2511 return;
2512 break;
2513 default:
2514 svdm_version = typec_get_negotiated_svdm_version(port->typec_port);
2515 if (svdm_version < 0)
2516 return;
2517 break;
2518 }
2519
2520 if (WARN_ON(count > VDO_MAX_SIZE - 1))
2521 count = VDO_MAX_SIZE - 1;
2522
2523 /* set VDM header with VID & CMD */
2524 header = VDO(vid, ((vid & USB_SID_PD) == USB_SID_PD) ?
2525 1 : (PD_VDO_CMD(cmd) <= CMD_ATTENTION),
2526 svdm_version, cmd);
2527 tcpm_queue_vdm(port, header, data, count, tx_sop_type);
2528 }
2529
vdm_ready_timeout(u32 vdm_hdr)2530 static unsigned int vdm_ready_timeout(u32 vdm_hdr)
2531 {
2532 unsigned int timeout;
2533 int cmd = PD_VDO_CMD(vdm_hdr);
2534
2535 /* its not a structured VDM command */
2536 if (!PD_VDO_SVDM(vdm_hdr))
2537 return PD_T_VDM_UNSTRUCTURED;
2538
2539 switch (PD_VDO_CMDT(vdm_hdr)) {
2540 case CMDT_INIT:
2541 if (cmd == CMD_ENTER_MODE || cmd == CMD_EXIT_MODE)
2542 timeout = PD_T_VDM_WAIT_MODE_E;
2543 else
2544 timeout = PD_T_VDM_SNDR_RSP;
2545 break;
2546 default:
2547 if (cmd == CMD_ENTER_MODE || cmd == CMD_EXIT_MODE)
2548 timeout = PD_T_VDM_E_MODE;
2549 else
2550 timeout = PD_T_VDM_RCVR_RSP;
2551 break;
2552 }
2553 return timeout;
2554 }
2555
vdm_run_state_machine(struct tcpm_port * port)2556 static void vdm_run_state_machine(struct tcpm_port *port)
2557 {
2558 struct pd_message msg;
2559 int i, res = 0;
2560 u32 vdo_hdr = port->vdo_data[0];
2561 u32 response[8] = { };
2562
2563 switch (port->vdm_state) {
2564 case VDM_STATE_READY:
2565 /* Only transmit VDM if attached */
2566 if (!port->attached) {
2567 port->vdm_state = VDM_STATE_ERR_BUSY;
2568 break;
2569 }
2570
2571 /*
2572 * if there's traffic or we're not in PDO ready state don't send
2573 * a VDM.
2574 */
2575 if (port->state != SRC_READY && port->state != SNK_READY &&
2576 port->state != SRC_VDM_IDENTITY_REQUEST) {
2577 port->vdm_sm_running = false;
2578 break;
2579 }
2580
2581 /* TODO: AMS operation for Unstructured VDM */
2582 if (PD_VDO_SVDM(vdo_hdr) && PD_VDO_CMDT(vdo_hdr) == CMDT_INIT) {
2583 switch (PD_VDO_CMD(vdo_hdr)) {
2584 case CMD_DISCOVER_IDENT:
2585 res = tcpm_ams_start(port, DISCOVER_IDENTITY);
2586 if (res == 0) {
2587 switch (port->tx_sop_type) {
2588 case TCPC_TX_SOP_PRIME:
2589 port->send_discover_prime = false;
2590 break;
2591 case TCPC_TX_SOP:
2592 port->send_discover = false;
2593 break;
2594 default:
2595 port->send_discover = false;
2596 break;
2597 }
2598 } else if (res == -EAGAIN) {
2599 port->vdo_data[0] = 0;
2600 mod_send_discover_delayed_work(port,
2601 SEND_DISCOVER_RETRY_MS);
2602 }
2603 break;
2604 case CMD_DISCOVER_SVID:
2605 res = tcpm_ams_start(port, DISCOVER_SVIDS);
2606 break;
2607 case CMD_DISCOVER_MODES:
2608 res = tcpm_ams_start(port, DISCOVER_MODES);
2609 break;
2610 case CMD_ENTER_MODE:
2611 res = tcpm_ams_start(port, DFP_TO_UFP_ENTER_MODE);
2612 break;
2613 case CMD_EXIT_MODE:
2614 res = tcpm_ams_start(port, DFP_TO_UFP_EXIT_MODE);
2615 break;
2616 case CMD_ATTENTION:
2617 res = tcpm_ams_start(port, ATTENTION);
2618 break;
2619 case VDO_CMD_VENDOR(0) ... VDO_CMD_VENDOR(15):
2620 res = tcpm_ams_start(port, STRUCTURED_VDMS);
2621 break;
2622 default:
2623 res = -EOPNOTSUPP;
2624 break;
2625 }
2626
2627 if (res < 0) {
2628 port->vdm_state = VDM_STATE_ERR_BUSY;
2629 return;
2630 }
2631 }
2632
2633 port->vdm_state = VDM_STATE_SEND_MESSAGE;
2634 mod_vdm_delayed_work(port, (port->negotiated_rev >= PD_REV30 &&
2635 port->pwr_role == TYPEC_SOURCE &&
2636 PD_VDO_SVDM(vdo_hdr) &&
2637 PD_VDO_CMDT(vdo_hdr) == CMDT_INIT) ?
2638 PD_T_SINK_TX : 0);
2639 break;
2640 case VDM_STATE_WAIT_RSP_BUSY:
2641 port->vdo_data[0] = port->vdo_retry;
2642 port->vdo_count = 1;
2643 port->vdm_state = VDM_STATE_READY;
2644 tcpm_ams_finish(port);
2645 break;
2646 case VDM_STATE_BUSY:
2647 port->vdm_state = VDM_STATE_ERR_TMOUT;
2648 if (port->ams != NONE_AMS)
2649 tcpm_ams_finish(port);
2650 break;
2651 case VDM_STATE_ERR_SEND:
2652 /*
2653 * When sending Discover Identity to SOP' before establishing an
2654 * explicit contract, do not retry. Instead, weave sending
2655 * Source_Capabilities over SOP and Discover Identity over SOP'.
2656 */
2657 if (port->state == SRC_VDM_IDENTITY_REQUEST) {
2658 tcpm_ams_finish(port);
2659 port->vdm_state = VDM_STATE_DONE;
2660 tcpm_set_state(port, SRC_SEND_CAPABILITIES, 0);
2661 /*
2662 * A partner which does not support USB PD will not reply,
2663 * so this is not a fatal error. At the same time, some
2664 * devices may not return GoodCRC under some circumstances,
2665 * so we need to retry.
2666 */
2667 } else if (port->vdm_retries < 3) {
2668 tcpm_log(port, "VDM Tx error, retry");
2669 port->vdm_retries++;
2670 port->vdm_state = VDM_STATE_READY;
2671 if (PD_VDO_SVDM(vdo_hdr) && PD_VDO_CMDT(vdo_hdr) == CMDT_INIT)
2672 tcpm_ams_finish(port);
2673 } else {
2674 tcpm_ams_finish(port);
2675 if (port->tx_sop_type == TCPC_TX_SOP)
2676 break;
2677 /* Handle SOP' Transmission Errors */
2678 switch (PD_VDO_CMD(vdo_hdr)) {
2679 /*
2680 * If Discover Identity fails on SOP', then resume
2681 * discovery process on SOP only.
2682 */
2683 case CMD_DISCOVER_IDENT:
2684 port->vdo_data[0] = 0;
2685 response[0] = VDO(USB_SID_PD, 1,
2686 typec_get_negotiated_svdm_version(
2687 port->typec_port),
2688 CMD_DISCOVER_SVID);
2689 tcpm_queue_vdm(port, response[0], &response[1],
2690 0, TCPC_TX_SOP);
2691 break;
2692 /*
2693 * If Discover SVIDs or Discover Modes fail, then
2694 * proceed with Alt Mode discovery process on SOP.
2695 */
2696 case CMD_DISCOVER_SVID:
2697 tcpm_register_partner_altmodes(port);
2698 break;
2699 case CMD_DISCOVER_MODES:
2700 tcpm_register_partner_altmodes(port);
2701 break;
2702 default:
2703 break;
2704 }
2705 }
2706 break;
2707 case VDM_STATE_SEND_MESSAGE:
2708 /* Prepare and send VDM */
2709 memset(&msg, 0, sizeof(msg));
2710 if (port->tx_sop_type == TCPC_TX_SOP_PRIME) {
2711 msg.header = PD_HEADER_LE(PD_DATA_VENDOR_DEF,
2712 0, /* Cable Plug Indicator for DFP/UFP */
2713 0, /* Reserved */
2714 port->negotiated_rev_prime,
2715 port->message_id_prime,
2716 port->vdo_count);
2717 } else {
2718 msg.header = PD_HEADER_LE(PD_DATA_VENDOR_DEF,
2719 port->pwr_role,
2720 port->data_role,
2721 port->negotiated_rev,
2722 port->message_id,
2723 port->vdo_count);
2724 }
2725 for (i = 0; i < port->vdo_count; i++)
2726 msg.payload[i] = cpu_to_le32(port->vdo_data[i]);
2727 res = tcpm_pd_transmit(port, port->tx_sop_type, &msg);
2728 if (res < 0) {
2729 port->vdm_state = VDM_STATE_ERR_SEND;
2730 } else {
2731 unsigned long timeout;
2732
2733 port->vdm_retries = 0;
2734 port->vdo_data[0] = 0;
2735 port->vdm_state = VDM_STATE_BUSY;
2736 timeout = vdm_ready_timeout(vdo_hdr);
2737 mod_vdm_delayed_work(port, timeout);
2738 }
2739 break;
2740 default:
2741 break;
2742 }
2743 }
2744
vdm_state_machine_work(struct kthread_work * work)2745 static void vdm_state_machine_work(struct kthread_work *work)
2746 {
2747 struct tcpm_port *port = container_of(work, struct tcpm_port, vdm_state_machine);
2748 enum vdm_states prev_state;
2749
2750 mutex_lock(&port->lock);
2751
2752 /*
2753 * Continue running as long as the port is not busy and there was
2754 * a state change.
2755 */
2756 do {
2757 prev_state = port->vdm_state;
2758 vdm_run_state_machine(port);
2759 } while (port->vdm_state != prev_state &&
2760 port->vdm_state != VDM_STATE_BUSY &&
2761 port->vdm_state != VDM_STATE_SEND_MESSAGE);
2762
2763 if (port->vdm_state < VDM_STATE_READY)
2764 port->vdm_sm_running = false;
2765
2766 mutex_unlock(&port->lock);
2767 }
2768
2769 enum pdo_err {
2770 PDO_NO_ERR,
2771 PDO_ERR_NO_VSAFE5V,
2772 PDO_ERR_VSAFE5V_NOT_FIRST,
2773 PDO_ERR_PDO_TYPE_NOT_IN_ORDER,
2774 PDO_ERR_FIXED_NOT_SORTED,
2775 PDO_ERR_VARIABLE_BATT_NOT_SORTED,
2776 PDO_ERR_DUPE_PDO,
2777 PDO_ERR_PPS_APDO_NOT_SORTED,
2778 PDO_ERR_DUPE_PPS_APDO,
2779 };
2780
2781 static const char * const pdo_err_msg[] = {
2782 [PDO_ERR_NO_VSAFE5V] =
2783 " err: source/sink caps should at least have vSafe5V",
2784 [PDO_ERR_VSAFE5V_NOT_FIRST] =
2785 " err: vSafe5V Fixed Supply Object Shall always be the first object",
2786 [PDO_ERR_PDO_TYPE_NOT_IN_ORDER] =
2787 " err: PDOs should be in the following order: Fixed; Battery; Variable",
2788 [PDO_ERR_FIXED_NOT_SORTED] =
2789 " err: Fixed supply pdos should be in increasing order of their fixed voltage",
2790 [PDO_ERR_VARIABLE_BATT_NOT_SORTED] =
2791 " err: Variable/Battery supply pdos should be in increasing order of their minimum voltage",
2792 [PDO_ERR_DUPE_PDO] =
2793 " err: Variable/Batt supply pdos cannot have same min/max voltage",
2794 [PDO_ERR_PPS_APDO_NOT_SORTED] =
2795 " err: Programmable power supply apdos should be in increasing order of their maximum voltage",
2796 [PDO_ERR_DUPE_PPS_APDO] =
2797 " err: Programmable power supply apdos cannot have same min/max voltage and max current",
2798 };
2799
tcpm_caps_err(struct tcpm_port * port,const u32 * pdo,unsigned int nr_pdo)2800 static enum pdo_err tcpm_caps_err(struct tcpm_port *port, const u32 *pdo,
2801 unsigned int nr_pdo)
2802 {
2803 unsigned int i;
2804
2805 /* Should at least contain vSafe5v */
2806 if (nr_pdo < 1)
2807 return PDO_ERR_NO_VSAFE5V;
2808
2809 /* The vSafe5V Fixed Supply Object Shall always be the first object */
2810 if (pdo_type(pdo[0]) != PDO_TYPE_FIXED ||
2811 pdo_fixed_voltage(pdo[0]) != VSAFE5V)
2812 return PDO_ERR_VSAFE5V_NOT_FIRST;
2813
2814 for (i = 1; i < nr_pdo; i++) {
2815 if (pdo_type(pdo[i]) < pdo_type(pdo[i - 1])) {
2816 return PDO_ERR_PDO_TYPE_NOT_IN_ORDER;
2817 } else if (pdo_type(pdo[i]) == pdo_type(pdo[i - 1])) {
2818 enum pd_pdo_type type = pdo_type(pdo[i]);
2819
2820 switch (type) {
2821 /*
2822 * The remaining Fixed Supply Objects, if
2823 * present, shall be sent in voltage order;
2824 * lowest to highest.
2825 */
2826 case PDO_TYPE_FIXED:
2827 if (pdo_fixed_voltage(pdo[i]) <=
2828 pdo_fixed_voltage(pdo[i - 1]))
2829 return PDO_ERR_FIXED_NOT_SORTED;
2830 break;
2831 /*
2832 * The Battery Supply Objects and Variable
2833 * supply, if present shall be sent in Minimum
2834 * Voltage order; lowest to highest.
2835 */
2836 case PDO_TYPE_VAR:
2837 case PDO_TYPE_BATT:
2838 if (pdo_min_voltage(pdo[i]) <
2839 pdo_min_voltage(pdo[i - 1]))
2840 return PDO_ERR_VARIABLE_BATT_NOT_SORTED;
2841 else if ((pdo_min_voltage(pdo[i]) ==
2842 pdo_min_voltage(pdo[i - 1])) &&
2843 (pdo_max_voltage(pdo[i]) ==
2844 pdo_max_voltage(pdo[i - 1])))
2845 return PDO_ERR_DUPE_PDO;
2846 break;
2847 /*
2848 * The Programmable Power Supply APDOs, if present,
2849 * shall be sent in Maximum Voltage order;
2850 * lowest to highest.
2851 */
2852 case PDO_TYPE_APDO:
2853 if (pdo_apdo_type(pdo[i]) != APDO_TYPE_PPS)
2854 break;
2855
2856 if (pdo_pps_apdo_max_voltage(pdo[i]) <
2857 pdo_pps_apdo_max_voltage(pdo[i - 1]))
2858 return PDO_ERR_PPS_APDO_NOT_SORTED;
2859 else if (pdo_pps_apdo_min_voltage(pdo[i]) ==
2860 pdo_pps_apdo_min_voltage(pdo[i - 1]) &&
2861 pdo_pps_apdo_max_voltage(pdo[i]) ==
2862 pdo_pps_apdo_max_voltage(pdo[i - 1]) &&
2863 pdo_pps_apdo_max_current(pdo[i]) ==
2864 pdo_pps_apdo_max_current(pdo[i - 1]))
2865 return PDO_ERR_DUPE_PPS_APDO;
2866 break;
2867 default:
2868 tcpm_log_force(port, " Unknown pdo type");
2869 }
2870 }
2871 }
2872
2873 return PDO_NO_ERR;
2874 }
2875
tcpm_validate_caps(struct tcpm_port * port,const u32 * pdo,unsigned int nr_pdo)2876 static int tcpm_validate_caps(struct tcpm_port *port, const u32 *pdo,
2877 unsigned int nr_pdo)
2878 {
2879 enum pdo_err err_index = tcpm_caps_err(port, pdo, nr_pdo);
2880
2881 if (err_index != PDO_NO_ERR) {
2882 tcpm_log_force(port, " %s", pdo_err_msg[err_index]);
2883 return -EINVAL;
2884 }
2885
2886 return 0;
2887 }
2888
tcpm_altmode_enter(struct typec_altmode * altmode,u32 * vdo)2889 static int tcpm_altmode_enter(struct typec_altmode *altmode, u32 *vdo)
2890 {
2891 struct tcpm_port *port = typec_altmode_get_drvdata(altmode);
2892 int svdm_version;
2893 u32 header;
2894
2895 svdm_version = typec_get_negotiated_svdm_version(port->typec_port);
2896 if (svdm_version < 0)
2897 return svdm_version;
2898
2899 header = VDO(altmode->svid, vdo ? 2 : 1, svdm_version, CMD_ENTER_MODE);
2900 header |= VDO_OPOS(altmode->mode);
2901
2902 return tcpm_queue_vdm_unlocked(port, header, vdo, vdo ? 1 : 0, TCPC_TX_SOP);
2903 }
2904
tcpm_altmode_exit(struct typec_altmode * altmode)2905 static int tcpm_altmode_exit(struct typec_altmode *altmode)
2906 {
2907 struct tcpm_port *port = typec_altmode_get_drvdata(altmode);
2908 int svdm_version;
2909 u32 header;
2910
2911 svdm_version = typec_get_negotiated_svdm_version(port->typec_port);
2912 if (svdm_version < 0)
2913 return svdm_version;
2914
2915 header = VDO(altmode->svid, 1, svdm_version, CMD_EXIT_MODE);
2916 header |= VDO_OPOS(altmode->mode);
2917
2918 return tcpm_queue_vdm_unlocked(port, header, NULL, 0, TCPC_TX_SOP);
2919 }
2920
tcpm_altmode_vdm(struct typec_altmode * altmode,u32 header,const u32 * data,int count)2921 static int tcpm_altmode_vdm(struct typec_altmode *altmode,
2922 u32 header, const u32 *data, int count)
2923 {
2924 struct tcpm_port *port = typec_altmode_get_drvdata(altmode);
2925
2926 return tcpm_queue_vdm_unlocked(port, header, data, count - 1, TCPC_TX_SOP);
2927 }
2928
2929 static const struct typec_altmode_ops tcpm_altmode_ops = {
2930 .enter = tcpm_altmode_enter,
2931 .exit = tcpm_altmode_exit,
2932 .vdm = tcpm_altmode_vdm,
2933 };
2934
2935
tcpm_cable_altmode_enter(struct typec_altmode * altmode,enum typec_plug_index sop,u32 * vdo)2936 static int tcpm_cable_altmode_enter(struct typec_altmode *altmode, enum typec_plug_index sop,
2937 u32 *vdo)
2938 {
2939 struct tcpm_port *port = typec_altmode_get_drvdata(altmode);
2940 int svdm_version;
2941 u32 header;
2942
2943 svdm_version = typec_get_cable_svdm_version(port->typec_port);
2944 if (svdm_version < 0)
2945 return svdm_version;
2946
2947 header = VDO(altmode->svid, vdo ? 2 : 1, svdm_version, CMD_ENTER_MODE);
2948 header |= VDO_OPOS(altmode->mode);
2949
2950 return tcpm_queue_vdm_unlocked(port, header, vdo, vdo ? 1 : 0, TCPC_TX_SOP_PRIME);
2951 }
2952
tcpm_cable_altmode_exit(struct typec_altmode * altmode,enum typec_plug_index sop)2953 static int tcpm_cable_altmode_exit(struct typec_altmode *altmode, enum typec_plug_index sop)
2954 {
2955 struct tcpm_port *port = typec_altmode_get_drvdata(altmode);
2956 int svdm_version;
2957 u32 header;
2958
2959 svdm_version = typec_get_cable_svdm_version(port->typec_port);
2960 if (svdm_version < 0)
2961 return svdm_version;
2962
2963 header = VDO(altmode->svid, 1, svdm_version, CMD_EXIT_MODE);
2964 header |= VDO_OPOS(altmode->mode);
2965
2966 return tcpm_queue_vdm_unlocked(port, header, NULL, 0, TCPC_TX_SOP_PRIME);
2967 }
2968
tcpm_cable_altmode_vdm(struct typec_altmode * altmode,enum typec_plug_index sop,u32 header,const u32 * data,int count)2969 static int tcpm_cable_altmode_vdm(struct typec_altmode *altmode, enum typec_plug_index sop,
2970 u32 header, const u32 *data, int count)
2971 {
2972 struct tcpm_port *port = typec_altmode_get_drvdata(altmode);
2973
2974 return tcpm_queue_vdm_unlocked(port, header, data, count - 1, TCPC_TX_SOP_PRIME);
2975 }
2976
2977 static const struct typec_cable_ops tcpm_cable_ops = {
2978 .enter = tcpm_cable_altmode_enter,
2979 .exit = tcpm_cable_altmode_exit,
2980 .vdm = tcpm_cable_altmode_vdm,
2981 };
2982
2983 /*
2984 * PD (data, control) command handling functions
2985 */
ready_state(struct tcpm_port * port)2986 static inline enum tcpm_state ready_state(struct tcpm_port *port)
2987 {
2988 if (port->pwr_role == TYPEC_SOURCE)
2989 return SRC_READY;
2990 else
2991 return SNK_READY;
2992 }
2993
2994 static int tcpm_pd_send_control(struct tcpm_port *port,
2995 enum pd_ctrl_msg_type type,
2996 enum tcpm_transmit_type tx_sop_type);
2997
tcpm_handle_alert(struct tcpm_port * port,const __le32 * payload,int cnt)2998 static void tcpm_handle_alert(struct tcpm_port *port, const __le32 *payload,
2999 int cnt)
3000 {
3001 u32 p0 = le32_to_cpu(payload[0]);
3002 unsigned int type = usb_pd_ado_type(p0);
3003
3004 if (!type) {
3005 tcpm_log(port, "Alert message received with no type");
3006 tcpm_queue_message(port, PD_MSG_CTRL_NOT_SUPP);
3007 return;
3008 }
3009
3010 /* Just handling non-battery alerts for now */
3011 if (!(type & USB_PD_ADO_TYPE_BATT_STATUS_CHANGE)) {
3012 if (port->pwr_role == TYPEC_SOURCE) {
3013 port->upcoming_state = GET_STATUS_SEND;
3014 tcpm_ams_start(port, GETTING_SOURCE_SINK_STATUS);
3015 } else {
3016 /*
3017 * Do not check SinkTxOk here in case the Source doesn't set its Rp to
3018 * SinkTxOk in time.
3019 */
3020 port->ams = GETTING_SOURCE_SINK_STATUS;
3021 tcpm_set_state(port, GET_STATUS_SEND, 0);
3022 }
3023 } else {
3024 tcpm_queue_message(port, PD_MSG_CTRL_NOT_SUPP);
3025 }
3026 }
3027
tcpm_set_auto_vbus_discharge_threshold(struct tcpm_port * port,enum typec_pwr_opmode mode,bool pps_active,u32 requested_vbus_voltage)3028 static int tcpm_set_auto_vbus_discharge_threshold(struct tcpm_port *port,
3029 enum typec_pwr_opmode mode, bool pps_active,
3030 u32 requested_vbus_voltage)
3031 {
3032 int ret;
3033
3034 if (!port->tcpc->set_auto_vbus_discharge_threshold)
3035 return 0;
3036
3037 ret = port->tcpc->set_auto_vbus_discharge_threshold(port->tcpc, mode, pps_active,
3038 requested_vbus_voltage,
3039 port->pps_data.min_volt);
3040 tcpm_log_force(port,
3041 "set_auto_vbus_discharge_threshold mode:%d pps_active:%c vbus:%u pps_apdo_min_volt:%u ret:%d",
3042 mode, pps_active ? 'y' : 'n', requested_vbus_voltage,
3043 port->pps_data.min_volt, ret);
3044
3045 return ret;
3046 }
3047
tcpm_pd_handle_state(struct tcpm_port * port,enum tcpm_state state,enum tcpm_ams ams,unsigned int delay_ms)3048 static void tcpm_pd_handle_state(struct tcpm_port *port,
3049 enum tcpm_state state,
3050 enum tcpm_ams ams,
3051 unsigned int delay_ms)
3052 {
3053 switch (port->state) {
3054 case SRC_READY:
3055 case SNK_READY:
3056 port->ams = ams;
3057 tcpm_set_state(port, state, delay_ms);
3058 break;
3059 /* 8.3.3.4.1.1 and 6.8.1 power transitioning */
3060 case SNK_TRANSITION_SINK:
3061 case SNK_TRANSITION_SINK_VBUS:
3062 case SRC_TRANSITION_SUPPLY:
3063 tcpm_set_state(port, HARD_RESET_SEND, 0);
3064 break;
3065 default:
3066 if (!tcpm_ams_interruptible(port)) {
3067 tcpm_set_state(port, port->pwr_role == TYPEC_SOURCE ?
3068 SRC_SOFT_RESET_WAIT_SNK_TX :
3069 SNK_SOFT_RESET,
3070 0);
3071 } else {
3072 /* process the Message 6.8.1 */
3073 port->upcoming_state = state;
3074 port->next_ams = ams;
3075 tcpm_set_state(port, ready_state(port), delay_ms);
3076 }
3077 break;
3078 }
3079 }
3080
tcpm_pd_handle_msg(struct tcpm_port * port,enum pd_msg_request message,enum tcpm_ams ams)3081 static void tcpm_pd_handle_msg(struct tcpm_port *port,
3082 enum pd_msg_request message,
3083 enum tcpm_ams ams)
3084 {
3085 switch (port->state) {
3086 case SRC_READY:
3087 case SNK_READY:
3088 port->ams = ams;
3089 tcpm_queue_message(port, message);
3090 break;
3091 /* PD 3.0 Spec 8.3.3.4.1.1 and 6.8.1 */
3092 case SNK_TRANSITION_SINK:
3093 case SNK_TRANSITION_SINK_VBUS:
3094 case SRC_TRANSITION_SUPPLY:
3095 tcpm_set_state(port, HARD_RESET_SEND, 0);
3096 break;
3097 default:
3098 if (!tcpm_ams_interruptible(port)) {
3099 tcpm_set_state(port, port->pwr_role == TYPEC_SOURCE ?
3100 SRC_SOFT_RESET_WAIT_SNK_TX :
3101 SNK_SOFT_RESET,
3102 0);
3103 } else {
3104 port->next_ams = ams;
3105 tcpm_set_state(port, ready_state(port), 0);
3106 /* 6.8.1 process the Message */
3107 tcpm_queue_message(port, message);
3108 }
3109 break;
3110 }
3111 }
3112
tcpm_register_source_caps(struct tcpm_port * port)3113 static int tcpm_register_source_caps(struct tcpm_port *port)
3114 {
3115 struct usb_power_delivery_desc desc = { port->negotiated_rev };
3116 struct usb_power_delivery_capabilities_desc caps = { };
3117 struct usb_power_delivery_capabilities *cap = port->partner_source_caps;
3118
3119 if (!port->partner_pd)
3120 port->partner_pd = usb_power_delivery_register(NULL, &desc);
3121 if (IS_ERR(port->partner_pd))
3122 return PTR_ERR(port->partner_pd);
3123
3124 memcpy(caps.pdo, port->source_caps, sizeof(u32) * port->nr_source_caps);
3125 caps.role = TYPEC_SOURCE;
3126
3127 if (cap) {
3128 usb_power_delivery_unregister_capabilities(cap);
3129 port->partner_source_caps = NULL;
3130 }
3131
3132 cap = usb_power_delivery_register_capabilities(port->partner_pd, &caps);
3133 if (IS_ERR(cap))
3134 return PTR_ERR(cap);
3135
3136 port->partner_source_caps = cap;
3137
3138 return 0;
3139 }
3140
tcpm_register_sink_caps(struct tcpm_port * port)3141 static int tcpm_register_sink_caps(struct tcpm_port *port)
3142 {
3143 struct usb_power_delivery_desc desc = { port->negotiated_rev };
3144 struct usb_power_delivery_capabilities_desc caps = { };
3145 struct usb_power_delivery_capabilities *cap;
3146
3147 if (!port->partner_pd)
3148 port->partner_pd = usb_power_delivery_register(NULL, &desc);
3149 if (IS_ERR(port->partner_pd))
3150 return PTR_ERR(port->partner_pd);
3151
3152 memcpy(caps.pdo, port->sink_caps, sizeof(u32) * port->nr_sink_caps);
3153 caps.role = TYPEC_SINK;
3154
3155 cap = usb_power_delivery_register_capabilities(port->partner_pd, &caps);
3156 if (IS_ERR(cap))
3157 return PTR_ERR(cap);
3158
3159 port->partner_sink_caps = cap;
3160
3161 return 0;
3162 }
3163
tcpm_pd_data_request(struct tcpm_port * port,const struct pd_message * msg,enum tcpm_transmit_type rx_sop_type)3164 static void tcpm_pd_data_request(struct tcpm_port *port,
3165 const struct pd_message *msg,
3166 enum tcpm_transmit_type rx_sop_type)
3167 {
3168 enum pd_data_msg_type type = pd_header_type_le(msg->header);
3169 unsigned int cnt = pd_header_cnt_le(msg->header);
3170 unsigned int rev = pd_header_rev_le(msg->header);
3171 unsigned int i;
3172 enum frs_typec_current partner_frs_current;
3173 bool frs_enable;
3174 int ret;
3175
3176 if (tcpm_vdm_ams(port) && type != PD_DATA_VENDOR_DEF) {
3177 port->vdm_state = VDM_STATE_ERR_BUSY;
3178 tcpm_ams_finish(port);
3179 mod_vdm_delayed_work(port, 0);
3180 }
3181
3182 switch (type) {
3183 case PD_DATA_SOURCE_CAP:
3184 for (i = 0; i < cnt; i++)
3185 port->source_caps[i] = le32_to_cpu(msg->payload[i]);
3186
3187 port->nr_source_caps = cnt;
3188
3189 tcpm_log_source_caps(port);
3190
3191 tcpm_validate_caps(port, port->source_caps,
3192 port->nr_source_caps);
3193
3194 tcpm_register_source_caps(port);
3195
3196 /*
3197 * Adjust revision in subsequent message headers, as required,
3198 * to comply with 6.2.1.1.5 of the USB PD 3.0 spec. We don't
3199 * support Rev 1.0 so just do nothing in that scenario.
3200 */
3201 if (rev == PD_REV10) {
3202 if (port->ams == GET_SOURCE_CAPABILITIES)
3203 tcpm_ams_finish(port);
3204 break;
3205 }
3206
3207 if (rev < PD_MAX_REV) {
3208 port->negotiated_rev = rev;
3209 if (port->negotiated_rev_prime > port->negotiated_rev)
3210 port->negotiated_rev_prime = port->negotiated_rev;
3211 }
3212
3213 if (port->pwr_role == TYPEC_SOURCE) {
3214 if (port->ams == GET_SOURCE_CAPABILITIES)
3215 tcpm_pd_handle_state(port, SRC_READY, NONE_AMS, 0);
3216 /* Unexpected Source Capabilities */
3217 else
3218 tcpm_pd_handle_msg(port,
3219 port->negotiated_rev < PD_REV30 ?
3220 PD_MSG_CTRL_REJECT :
3221 PD_MSG_CTRL_NOT_SUPP,
3222 NONE_AMS);
3223 } else if (port->state == SNK_WAIT_CAPABILITIES ||
3224 port->state == SNK_WAIT_CAPABILITIES_TIMEOUT) {
3225 /*
3226 * This message may be received even if VBUS is not
3227 * present. This is quite unexpected; see USB PD
3228 * specification, sections 8.3.3.6.3.1 and 8.3.3.6.3.2.
3229 * However, at the same time, we must be ready to
3230 * receive this message and respond to it 15ms after
3231 * receiving PS_RDY during power swap operations, no matter
3232 * if VBUS is available or not (USB PD specification,
3233 * section 6.5.9.2).
3234 * So we need to accept the message either way,
3235 * but be prepared to keep waiting for VBUS after it was
3236 * handled.
3237 */
3238 port->ams = POWER_NEGOTIATION;
3239 port->in_ams = true;
3240 tcpm_set_state(port, SNK_NEGOTIATE_CAPABILITIES, 0);
3241 } else {
3242 if (port->ams == GET_SOURCE_CAPABILITIES)
3243 tcpm_ams_finish(port);
3244 tcpm_pd_handle_state(port, SNK_NEGOTIATE_CAPABILITIES,
3245 POWER_NEGOTIATION, 0);
3246 }
3247 break;
3248 case PD_DATA_REQUEST:
3249 /*
3250 * Adjust revision in subsequent message headers, as required,
3251 * to comply with 6.2.1.1.5 of the USB PD 3.0 spec. We don't
3252 * support Rev 1.0 so just reject in that scenario.
3253 */
3254 if (rev == PD_REV10) {
3255 tcpm_pd_handle_msg(port,
3256 port->negotiated_rev < PD_REV30 ?
3257 PD_MSG_CTRL_REJECT :
3258 PD_MSG_CTRL_NOT_SUPP,
3259 NONE_AMS);
3260 break;
3261 }
3262
3263 if (rev < PD_MAX_REV) {
3264 port->negotiated_rev = rev;
3265 if (port->negotiated_rev_prime > port->negotiated_rev)
3266 port->negotiated_rev_prime = port->negotiated_rev;
3267 }
3268
3269 if (port->pwr_role != TYPEC_SOURCE || cnt != 1) {
3270 tcpm_pd_handle_msg(port,
3271 port->negotiated_rev < PD_REV30 ?
3272 PD_MSG_CTRL_REJECT :
3273 PD_MSG_CTRL_NOT_SUPP,
3274 NONE_AMS);
3275 break;
3276 }
3277
3278 port->sink_request = le32_to_cpu(msg->payload[0]);
3279
3280 if (port->vdm_sm_running && port->explicit_contract) {
3281 tcpm_pd_handle_msg(port, PD_MSG_CTRL_WAIT, port->ams);
3282 break;
3283 }
3284
3285 if (port->state == SRC_SEND_CAPABILITIES)
3286 tcpm_set_state(port, SRC_NEGOTIATE_CAPABILITIES, 0);
3287 else
3288 tcpm_pd_handle_state(port, SRC_NEGOTIATE_CAPABILITIES,
3289 POWER_NEGOTIATION, 0);
3290 break;
3291 case PD_DATA_SINK_CAP:
3292 /* We don't do anything with this at the moment... */
3293 for (i = 0; i < cnt; i++)
3294 port->sink_caps[i] = le32_to_cpu(msg->payload[i]);
3295
3296 partner_frs_current = (port->sink_caps[0] & PDO_FIXED_FRS_CURR_MASK) >>
3297 PDO_FIXED_FRS_CURR_SHIFT;
3298 frs_enable = partner_frs_current && (partner_frs_current <=
3299 port->new_source_frs_current);
3300 tcpm_log(port,
3301 "Port partner FRS capable partner_frs_current:%u port_frs_current:%u enable:%c",
3302 partner_frs_current, port->new_source_frs_current, frs_enable ? 'y' : 'n');
3303 if (frs_enable) {
3304 ret = port->tcpc->enable_frs(port->tcpc, true);
3305 tcpm_log(port, "Enable FRS %s, ret:%d\n", ret ? "fail" : "success", ret);
3306 }
3307
3308 port->nr_sink_caps = cnt;
3309 port->sink_cap_done = true;
3310 tcpm_register_sink_caps(port);
3311
3312 if (port->ams == GET_SINK_CAPABILITIES)
3313 tcpm_set_state(port, ready_state(port), 0);
3314 /* Unexpected Sink Capabilities */
3315 else
3316 tcpm_pd_handle_msg(port,
3317 port->negotiated_rev < PD_REV30 ?
3318 PD_MSG_CTRL_REJECT :
3319 PD_MSG_CTRL_NOT_SUPP,
3320 NONE_AMS);
3321 break;
3322 case PD_DATA_VENDOR_DEF:
3323 tcpm_handle_vdm_request(port, msg->payload, cnt, rx_sop_type);
3324 break;
3325 case PD_DATA_BIST:
3326 port->bist_request = le32_to_cpu(msg->payload[0]);
3327 tcpm_pd_handle_state(port, BIST_RX, BIST, 0);
3328 break;
3329 case PD_DATA_ALERT:
3330 if (port->state != SRC_READY && port->state != SNK_READY)
3331 tcpm_pd_handle_state(port, port->pwr_role == TYPEC_SOURCE ?
3332 SRC_SOFT_RESET_WAIT_SNK_TX : SNK_SOFT_RESET,
3333 NONE_AMS, 0);
3334 else
3335 tcpm_handle_alert(port, msg->payload, cnt);
3336 break;
3337 case PD_DATA_BATT_STATUS:
3338 case PD_DATA_GET_COUNTRY_INFO:
3339 /* Currently unsupported */
3340 tcpm_pd_handle_msg(port, port->negotiated_rev < PD_REV30 ?
3341 PD_MSG_CTRL_REJECT :
3342 PD_MSG_CTRL_NOT_SUPP,
3343 NONE_AMS);
3344 break;
3345 default:
3346 tcpm_pd_handle_msg(port, port->negotiated_rev < PD_REV30 ?
3347 PD_MSG_CTRL_REJECT :
3348 PD_MSG_CTRL_NOT_SUPP,
3349 NONE_AMS);
3350 tcpm_log(port, "Unrecognized data message type %#x", type);
3351 break;
3352 }
3353 }
3354
tcpm_pps_complete(struct tcpm_port * port,int result)3355 static void tcpm_pps_complete(struct tcpm_port *port, int result)
3356 {
3357 if (port->pps_pending) {
3358 port->pps_status = result;
3359 port->pps_pending = false;
3360 complete(&port->pps_complete);
3361 }
3362 }
3363
tcpm_pd_ctrl_request(struct tcpm_port * port,const struct pd_message * msg,enum tcpm_transmit_type rx_sop_type)3364 static void tcpm_pd_ctrl_request(struct tcpm_port *port,
3365 const struct pd_message *msg,
3366 enum tcpm_transmit_type rx_sop_type)
3367 {
3368 enum pd_ctrl_msg_type type = pd_header_type_le(msg->header);
3369 enum tcpm_state next_state;
3370 unsigned int rev = pd_header_rev_le(msg->header);
3371
3372 /*
3373 * Stop VDM state machine if interrupted by other Messages while NOT_SUPP is allowed in
3374 * VDM AMS if waiting for VDM responses and will be handled later.
3375 */
3376 if (tcpm_vdm_ams(port) && type != PD_CTRL_NOT_SUPP && type != PD_CTRL_GOOD_CRC) {
3377 port->vdm_state = VDM_STATE_ERR_BUSY;
3378 tcpm_ams_finish(port);
3379 mod_vdm_delayed_work(port, 0);
3380 }
3381
3382 switch (type) {
3383 case PD_CTRL_GOOD_CRC:
3384 case PD_CTRL_PING:
3385 break;
3386 case PD_CTRL_GET_SOURCE_CAP:
3387 tcpm_pd_handle_msg(port, PD_MSG_DATA_SOURCE_CAP, GET_SOURCE_CAPABILITIES);
3388 break;
3389 case PD_CTRL_GET_SINK_CAP:
3390 tcpm_pd_handle_msg(port, PD_MSG_DATA_SINK_CAP, GET_SINK_CAPABILITIES);
3391 break;
3392 case PD_CTRL_GOTO_MIN:
3393 break;
3394 case PD_CTRL_PS_RDY:
3395 switch (port->state) {
3396 case SNK_TRANSITION_SINK:
3397 if (port->vbus_present) {
3398 tcpm_set_current_limit(port,
3399 port->req_current_limit,
3400 port->req_supply_voltage);
3401 port->explicit_contract = true;
3402 tcpm_set_auto_vbus_discharge_threshold(port,
3403 TYPEC_PWR_MODE_PD,
3404 port->pps_data.active,
3405 port->supply_voltage);
3406 tcpm_set_state(port, SNK_READY, 0);
3407 } else {
3408 /*
3409 * Seen after power swap. Keep waiting for VBUS
3410 * in a transitional state.
3411 */
3412 tcpm_set_state(port,
3413 SNK_TRANSITION_SINK_VBUS, 0);
3414 }
3415 break;
3416 case PR_SWAP_SRC_SNK_SOURCE_OFF_CC_DEBOUNCED:
3417 tcpm_set_state(port, PR_SWAP_SRC_SNK_SINK_ON, 0);
3418 break;
3419 case PR_SWAP_SNK_SRC_SINK_OFF:
3420 tcpm_set_state(port, PR_SWAP_SNK_SRC_SOURCE_ON, 0);
3421 break;
3422 case VCONN_SWAP_WAIT_FOR_VCONN:
3423 tcpm_set_state(port, VCONN_SWAP_TURN_OFF_VCONN, 0);
3424 break;
3425 case FR_SWAP_SNK_SRC_TRANSITION_TO_OFF:
3426 tcpm_set_state(port, FR_SWAP_SNK_SRC_NEW_SINK_READY, 0);
3427 break;
3428 default:
3429 tcpm_pd_handle_state(port,
3430 port->pwr_role == TYPEC_SOURCE ?
3431 SRC_SOFT_RESET_WAIT_SNK_TX :
3432 SNK_SOFT_RESET,
3433 NONE_AMS, 0);
3434 break;
3435 }
3436 break;
3437 case PD_CTRL_REJECT:
3438 case PD_CTRL_WAIT:
3439 case PD_CTRL_NOT_SUPP:
3440 switch (port->state) {
3441 case SNK_NEGOTIATE_CAPABILITIES:
3442 /* USB PD specification, Figure 8-43 */
3443 if (port->explicit_contract)
3444 next_state = SNK_READY;
3445 else
3446 next_state = SNK_WAIT_CAPABILITIES;
3447
3448 /* Threshold was relaxed before sending Request. Restore it back. */
3449 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_PD,
3450 port->pps_data.active,
3451 port->supply_voltage);
3452 tcpm_set_state(port, next_state, 0);
3453 break;
3454 case SNK_NEGOTIATE_PPS_CAPABILITIES:
3455 /* Revert data back from any requested PPS updates */
3456 port->pps_data.req_out_volt = port->supply_voltage;
3457 port->pps_data.req_op_curr = port->current_limit;
3458 port->pps_status = (type == PD_CTRL_WAIT ?
3459 -EAGAIN : -EOPNOTSUPP);
3460
3461 /* Threshold was relaxed before sending Request. Restore it back. */
3462 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_PD,
3463 port->pps_data.active,
3464 port->supply_voltage);
3465
3466 tcpm_set_state(port, SNK_READY, 0);
3467 break;
3468 case DR_SWAP_SEND:
3469 port->swap_status = (type == PD_CTRL_WAIT ?
3470 -EAGAIN : -EOPNOTSUPP);
3471 tcpm_set_state(port, DR_SWAP_CANCEL, 0);
3472 break;
3473 case PR_SWAP_SEND:
3474 port->swap_status = (type == PD_CTRL_WAIT ?
3475 -EAGAIN : -EOPNOTSUPP);
3476 tcpm_set_state(port, PR_SWAP_CANCEL, 0);
3477 break;
3478 case VCONN_SWAP_SEND:
3479 port->swap_status = (type == PD_CTRL_WAIT ?
3480 -EAGAIN : -EOPNOTSUPP);
3481 tcpm_set_state(port, VCONN_SWAP_CANCEL, 0);
3482 break;
3483 case FR_SWAP_SEND:
3484 tcpm_set_state(port, FR_SWAP_CANCEL, 0);
3485 break;
3486 case GET_SINK_CAP:
3487 port->sink_cap_done = true;
3488 tcpm_set_state(port, ready_state(port), 0);
3489 break;
3490 /*
3491 * Some port partners do not support GET_STATUS, avoid soft reset the link to
3492 * prevent redundant power re-negotiation
3493 */
3494 case GET_STATUS_SEND:
3495 tcpm_set_state(port, ready_state(port), 0);
3496 break;
3497 case SRC_READY:
3498 case SNK_READY:
3499 if (port->vdm_state > VDM_STATE_READY) {
3500 port->vdm_state = VDM_STATE_DONE;
3501 if (tcpm_vdm_ams(port))
3502 tcpm_ams_finish(port);
3503 mod_vdm_delayed_work(port, 0);
3504 break;
3505 }
3506 fallthrough;
3507 default:
3508 tcpm_pd_handle_state(port,
3509 port->pwr_role == TYPEC_SOURCE ?
3510 SRC_SOFT_RESET_WAIT_SNK_TX :
3511 SNK_SOFT_RESET,
3512 NONE_AMS, 0);
3513 break;
3514 }
3515 break;
3516 case PD_CTRL_ACCEPT:
3517 switch (port->state) {
3518 case SNK_NEGOTIATE_CAPABILITIES:
3519 port->pps_data.active = false;
3520 tcpm_set_state(port, SNK_TRANSITION_SINK, 0);
3521 break;
3522 case SNK_NEGOTIATE_PPS_CAPABILITIES:
3523 port->pps_data.active = true;
3524 port->pps_data.min_volt = port->pps_data.req_min_volt;
3525 port->pps_data.max_volt = port->pps_data.req_max_volt;
3526 port->pps_data.max_curr = port->pps_data.req_max_curr;
3527 port->req_supply_voltage = port->pps_data.req_out_volt;
3528 port->req_current_limit = port->pps_data.req_op_curr;
3529 power_supply_changed(port->psy);
3530 tcpm_set_state(port, SNK_TRANSITION_SINK, 0);
3531 break;
3532 case SOFT_RESET_SEND:
3533 if (port->ams == SOFT_RESET_AMS)
3534 tcpm_ams_finish(port);
3535 /*
3536 * SOP' Soft Reset is done after Vconn Swap,
3537 * which returns to ready state
3538 */
3539 if (rx_sop_type == TCPC_TX_SOP_PRIME) {
3540 if (rev < port->negotiated_rev_prime)
3541 port->negotiated_rev_prime = rev;
3542 tcpm_set_state(port, ready_state(port), 0);
3543 break;
3544 }
3545 if (port->pwr_role == TYPEC_SOURCE) {
3546 port->upcoming_state = SRC_SEND_CAPABILITIES;
3547 tcpm_ams_start(port, POWER_NEGOTIATION);
3548 } else {
3549 tcpm_set_state(port, SNK_WAIT_CAPABILITIES, 0);
3550 }
3551 break;
3552 case DR_SWAP_SEND:
3553 tcpm_set_state(port, DR_SWAP_CHANGE_DR, 0);
3554 break;
3555 case PR_SWAP_SEND:
3556 tcpm_set_state(port, PR_SWAP_START, 0);
3557 break;
3558 case VCONN_SWAP_SEND:
3559 tcpm_set_state(port, VCONN_SWAP_START, 0);
3560 break;
3561 case FR_SWAP_SEND:
3562 tcpm_set_state(port, FR_SWAP_SNK_SRC_TRANSITION_TO_OFF, 0);
3563 break;
3564 default:
3565 tcpm_pd_handle_state(port,
3566 port->pwr_role == TYPEC_SOURCE ?
3567 SRC_SOFT_RESET_WAIT_SNK_TX :
3568 SNK_SOFT_RESET,
3569 NONE_AMS, 0);
3570 break;
3571 }
3572 break;
3573 case PD_CTRL_SOFT_RESET:
3574 port->ams = SOFT_RESET_AMS;
3575 tcpm_set_state(port, SOFT_RESET, 0);
3576 break;
3577 case PD_CTRL_DR_SWAP:
3578 /*
3579 * XXX
3580 * 6.3.9: If an alternate mode is active, a request to swap
3581 * alternate modes shall trigger a port reset.
3582 */
3583 if (port->typec_caps.data != TYPEC_PORT_DRD) {
3584 tcpm_pd_handle_msg(port,
3585 port->negotiated_rev < PD_REV30 ?
3586 PD_MSG_CTRL_REJECT :
3587 PD_MSG_CTRL_NOT_SUPP,
3588 NONE_AMS);
3589 } else {
3590 if (port->send_discover && port->negotiated_rev < PD_REV30) {
3591 tcpm_queue_message(port, PD_MSG_CTRL_WAIT);
3592 break;
3593 }
3594
3595 tcpm_pd_handle_state(port, DR_SWAP_ACCEPT, DATA_ROLE_SWAP, 0);
3596 }
3597 break;
3598 case PD_CTRL_PR_SWAP:
3599 if (port->port_type != TYPEC_PORT_DRP) {
3600 tcpm_pd_handle_msg(port,
3601 port->negotiated_rev < PD_REV30 ?
3602 PD_MSG_CTRL_REJECT :
3603 PD_MSG_CTRL_NOT_SUPP,
3604 NONE_AMS);
3605 } else {
3606 if (port->send_discover && port->negotiated_rev < PD_REV30) {
3607 tcpm_queue_message(port, PD_MSG_CTRL_WAIT);
3608 break;
3609 }
3610
3611 tcpm_pd_handle_state(port, PR_SWAP_ACCEPT, POWER_ROLE_SWAP, 0);
3612 }
3613 break;
3614 case PD_CTRL_VCONN_SWAP:
3615 if (port->send_discover && port->negotiated_rev < PD_REV30) {
3616 tcpm_queue_message(port, PD_MSG_CTRL_WAIT);
3617 break;
3618 }
3619
3620 tcpm_pd_handle_state(port, VCONN_SWAP_ACCEPT, VCONN_SWAP, 0);
3621 break;
3622 case PD_CTRL_GET_SOURCE_CAP_EXT:
3623 case PD_CTRL_GET_STATUS:
3624 case PD_CTRL_FR_SWAP:
3625 case PD_CTRL_GET_PPS_STATUS:
3626 case PD_CTRL_GET_COUNTRY_CODES:
3627 /* Currently not supported */
3628 tcpm_pd_handle_msg(port,
3629 port->negotiated_rev < PD_REV30 ?
3630 PD_MSG_CTRL_REJECT :
3631 PD_MSG_CTRL_NOT_SUPP,
3632 NONE_AMS);
3633 break;
3634 case PD_CTRL_GET_REVISION:
3635 if (port->negotiated_rev >= PD_REV30 && port->pd_rev.rev_major)
3636 tcpm_pd_handle_msg(port, PD_MSG_DATA_REV,
3637 REVISION_INFORMATION);
3638 else
3639 tcpm_pd_handle_msg(port,
3640 port->negotiated_rev < PD_REV30 ?
3641 PD_MSG_CTRL_REJECT :
3642 PD_MSG_CTRL_NOT_SUPP,
3643 NONE_AMS);
3644 break;
3645 default:
3646 tcpm_pd_handle_msg(port,
3647 port->negotiated_rev < PD_REV30 ?
3648 PD_MSG_CTRL_REJECT :
3649 PD_MSG_CTRL_NOT_SUPP,
3650 NONE_AMS);
3651 tcpm_log(port, "Unrecognized ctrl message type %#x", type);
3652 break;
3653 }
3654 }
3655
tcpm_pd_ext_msg_request(struct tcpm_port * port,const struct pd_message * msg)3656 static void tcpm_pd_ext_msg_request(struct tcpm_port *port,
3657 const struct pd_message *msg)
3658 {
3659 enum pd_ext_msg_type type = pd_header_type_le(msg->header);
3660 unsigned int data_size = pd_ext_header_data_size_le(msg->ext_msg.header);
3661
3662 /* stopping VDM state machine if interrupted by other Messages */
3663 if (tcpm_vdm_ams(port)) {
3664 port->vdm_state = VDM_STATE_ERR_BUSY;
3665 tcpm_ams_finish(port);
3666 mod_vdm_delayed_work(port, 0);
3667 }
3668
3669 if (!(le16_to_cpu(msg->ext_msg.header) & PD_EXT_HDR_CHUNKED)) {
3670 tcpm_pd_handle_msg(port, PD_MSG_CTRL_NOT_SUPP, NONE_AMS);
3671 tcpm_log(port, "Unchunked extended messages unsupported");
3672 return;
3673 }
3674
3675 if (data_size > PD_EXT_MAX_CHUNK_DATA) {
3676 tcpm_pd_handle_state(port, CHUNK_NOT_SUPP, NONE_AMS, PD_T_CHUNK_NOT_SUPP);
3677 tcpm_log(port, "Chunk handling not yet supported");
3678 return;
3679 }
3680
3681 switch (type) {
3682 case PD_EXT_STATUS:
3683 case PD_EXT_PPS_STATUS:
3684 if (port->ams == GETTING_SOURCE_SINK_STATUS) {
3685 tcpm_ams_finish(port);
3686 tcpm_set_state(port, ready_state(port), 0);
3687 } else {
3688 /* unexpected Status or PPS_Status Message */
3689 tcpm_pd_handle_state(port, port->pwr_role == TYPEC_SOURCE ?
3690 SRC_SOFT_RESET_WAIT_SNK_TX : SNK_SOFT_RESET,
3691 NONE_AMS, 0);
3692 }
3693 break;
3694 case PD_EXT_SOURCE_CAP_EXT:
3695 case PD_EXT_GET_BATT_CAP:
3696 case PD_EXT_GET_BATT_STATUS:
3697 case PD_EXT_BATT_CAP:
3698 case PD_EXT_GET_MANUFACTURER_INFO:
3699 case PD_EXT_MANUFACTURER_INFO:
3700 case PD_EXT_SECURITY_REQUEST:
3701 case PD_EXT_SECURITY_RESPONSE:
3702 case PD_EXT_FW_UPDATE_REQUEST:
3703 case PD_EXT_FW_UPDATE_RESPONSE:
3704 case PD_EXT_COUNTRY_INFO:
3705 case PD_EXT_COUNTRY_CODES:
3706 tcpm_pd_handle_msg(port, PD_MSG_CTRL_NOT_SUPP, NONE_AMS);
3707 break;
3708 default:
3709 tcpm_pd_handle_msg(port, PD_MSG_CTRL_NOT_SUPP, NONE_AMS);
3710 tcpm_log(port, "Unrecognized extended message type %#x", type);
3711 break;
3712 }
3713 }
3714
tcpm_pd_rx_handler(struct kthread_work * work)3715 static void tcpm_pd_rx_handler(struct kthread_work *work)
3716 {
3717 struct pd_rx_event *event = container_of(work,
3718 struct pd_rx_event, work);
3719 const struct pd_message *msg = &event->msg;
3720 unsigned int cnt = pd_header_cnt_le(msg->header);
3721 struct tcpm_port *port = event->port;
3722 enum tcpm_transmit_type rx_sop_type = event->rx_sop_type;
3723
3724 mutex_lock(&port->lock);
3725
3726 tcpm_log(port, "PD RX, header: %#x [%d]", le16_to_cpu(msg->header),
3727 port->attached);
3728
3729 if (port->attached) {
3730 enum pd_ctrl_msg_type type = pd_header_type_le(msg->header);
3731 unsigned int msgid = pd_header_msgid_le(msg->header);
3732
3733 /*
3734 * Drop SOP' messages if cannot receive via
3735 * tcpm_can_communicate_sop_prime
3736 */
3737 if (rx_sop_type == TCPC_TX_SOP_PRIME &&
3738 !tcpm_can_communicate_sop_prime(port))
3739 goto done;
3740
3741 /*
3742 * USB PD standard, 6.6.1.2:
3743 * "... if MessageID value in a received Message is the
3744 * same as the stored value, the receiver shall return a
3745 * GoodCRC Message with that MessageID value and drop
3746 * the Message (this is a retry of an already received
3747 * Message). Note: this shall not apply to the Soft_Reset
3748 * Message which always has a MessageID value of zero."
3749 */
3750 switch (rx_sop_type) {
3751 case TCPC_TX_SOP_PRIME:
3752 if (msgid == port->rx_msgid_prime)
3753 goto done;
3754 port->rx_msgid_prime = msgid;
3755 break;
3756 case TCPC_TX_SOP:
3757 default:
3758 if (msgid == port->rx_msgid && type != PD_CTRL_SOFT_RESET)
3759 goto done;
3760 port->rx_msgid = msgid;
3761 break;
3762 }
3763
3764 /*
3765 * If both ends believe to be DFP/host, we have a data role
3766 * mismatch.
3767 */
3768 if (!!(le16_to_cpu(msg->header) & PD_HEADER_DATA_ROLE) ==
3769 (port->data_role == TYPEC_HOST) && rx_sop_type == TCPC_TX_SOP) {
3770 tcpm_log(port,
3771 "Data role mismatch, initiating error recovery");
3772 tcpm_set_state(port, ERROR_RECOVERY, 0);
3773 } else {
3774 if (le16_to_cpu(msg->header) & PD_HEADER_EXT_HDR)
3775 tcpm_pd_ext_msg_request(port, msg);
3776 else if (cnt)
3777 tcpm_pd_data_request(port, msg, rx_sop_type);
3778 else
3779 tcpm_pd_ctrl_request(port, msg, rx_sop_type);
3780 }
3781 }
3782
3783 done:
3784 mutex_unlock(&port->lock);
3785 kfree(event);
3786 }
3787
tcpm_pd_receive(struct tcpm_port * port,const struct pd_message * msg,enum tcpm_transmit_type rx_sop_type)3788 void tcpm_pd_receive(struct tcpm_port *port, const struct pd_message *msg,
3789 enum tcpm_transmit_type rx_sop_type)
3790 {
3791 struct pd_rx_event *event;
3792
3793 event = kzalloc(sizeof(*event), GFP_ATOMIC);
3794 if (!event)
3795 return;
3796
3797 kthread_init_work(&event->work, tcpm_pd_rx_handler);
3798 event->port = port;
3799 event->rx_sop_type = rx_sop_type;
3800 memcpy(&event->msg, msg, sizeof(*msg));
3801 kthread_queue_work(port->wq, &event->work);
3802 }
3803 EXPORT_SYMBOL_GPL(tcpm_pd_receive);
3804
tcpm_pd_send_control(struct tcpm_port * port,enum pd_ctrl_msg_type type,enum tcpm_transmit_type tx_sop_type)3805 static int tcpm_pd_send_control(struct tcpm_port *port,
3806 enum pd_ctrl_msg_type type,
3807 enum tcpm_transmit_type tx_sop_type)
3808 {
3809 struct pd_message msg;
3810
3811 memset(&msg, 0, sizeof(msg));
3812 switch (tx_sop_type) {
3813 case TCPC_TX_SOP_PRIME:
3814 msg.header = PD_HEADER_LE(type,
3815 0, /* Cable Plug Indicator for DFP/UFP */
3816 0, /* Reserved */
3817 port->negotiated_rev,
3818 port->message_id_prime,
3819 0);
3820 break;
3821 case TCPC_TX_SOP:
3822 msg.header = PD_HEADER_LE(type,
3823 port->pwr_role,
3824 port->data_role,
3825 port->negotiated_rev,
3826 port->message_id,
3827 0);
3828 break;
3829 default:
3830 msg.header = PD_HEADER_LE(type,
3831 port->pwr_role,
3832 port->data_role,
3833 port->negotiated_rev,
3834 port->message_id,
3835 0);
3836 break;
3837 }
3838
3839 return tcpm_pd_transmit(port, tx_sop_type, &msg);
3840 }
3841
3842 /*
3843 * Send queued message without affecting state.
3844 * Return true if state machine should go back to sleep,
3845 * false otherwise.
3846 */
tcpm_send_queued_message(struct tcpm_port * port)3847 static bool tcpm_send_queued_message(struct tcpm_port *port)
3848 {
3849 enum pd_msg_request queued_message;
3850 int ret;
3851
3852 do {
3853 queued_message = port->queued_message;
3854 port->queued_message = PD_MSG_NONE;
3855
3856 switch (queued_message) {
3857 case PD_MSG_CTRL_WAIT:
3858 tcpm_pd_send_control(port, PD_CTRL_WAIT, TCPC_TX_SOP);
3859 break;
3860 case PD_MSG_CTRL_REJECT:
3861 tcpm_pd_send_control(port, PD_CTRL_REJECT, TCPC_TX_SOP);
3862 break;
3863 case PD_MSG_CTRL_NOT_SUPP:
3864 tcpm_pd_send_control(port, PD_CTRL_NOT_SUPP, TCPC_TX_SOP);
3865 break;
3866 case PD_MSG_DATA_SINK_CAP:
3867 ret = tcpm_pd_send_sink_caps(port);
3868 if (ret < 0) {
3869 tcpm_log(port, "Unable to send snk caps, ret=%d", ret);
3870 tcpm_set_state(port, SNK_SOFT_RESET, 0);
3871 }
3872 tcpm_ams_finish(port);
3873 break;
3874 case PD_MSG_DATA_SOURCE_CAP:
3875 ret = tcpm_pd_send_source_caps(port);
3876 if (ret < 0) {
3877 tcpm_log(port,
3878 "Unable to send src caps, ret=%d",
3879 ret);
3880 tcpm_set_state(port, SOFT_RESET_SEND, 0);
3881 } else if (port->pwr_role == TYPEC_SOURCE) {
3882 tcpm_ams_finish(port);
3883 tcpm_set_state(port, HARD_RESET_SEND,
3884 PD_T_SENDER_RESPONSE);
3885 } else {
3886 tcpm_ams_finish(port);
3887 }
3888 break;
3889 case PD_MSG_DATA_REV:
3890 ret = tcpm_pd_send_revision(port);
3891 if (ret)
3892 tcpm_log(port,
3893 "Unable to send revision msg, ret=%d",
3894 ret);
3895 tcpm_ams_finish(port);
3896 break;
3897 default:
3898 break;
3899 }
3900 } while (port->queued_message != PD_MSG_NONE);
3901
3902 if (port->delayed_state != INVALID_STATE) {
3903 if (ktime_after(port->delayed_runtime, ktime_get())) {
3904 mod_tcpm_delayed_work(port, ktime_to_ms(ktime_sub(port->delayed_runtime,
3905 ktime_get())));
3906 return true;
3907 }
3908 port->delayed_state = INVALID_STATE;
3909 }
3910 return false;
3911 }
3912
tcpm_pd_check_request(struct tcpm_port * port)3913 static int tcpm_pd_check_request(struct tcpm_port *port)
3914 {
3915 u32 pdo, rdo = port->sink_request;
3916 unsigned int max, op, pdo_max, index;
3917 enum pd_pdo_type type;
3918
3919 index = rdo_index(rdo);
3920 if (!index || index > port->nr_src_pdo)
3921 return -EINVAL;
3922
3923 pdo = port->src_pdo[index - 1];
3924 type = pdo_type(pdo);
3925 switch (type) {
3926 case PDO_TYPE_FIXED:
3927 case PDO_TYPE_VAR:
3928 max = rdo_max_current(rdo);
3929 op = rdo_op_current(rdo);
3930 pdo_max = pdo_max_current(pdo);
3931
3932 if (op > pdo_max)
3933 return -EINVAL;
3934 if (max > pdo_max && !(rdo & RDO_CAP_MISMATCH))
3935 return -EINVAL;
3936
3937 if (type == PDO_TYPE_FIXED)
3938 tcpm_log(port,
3939 "Requested %u mV, %u mA for %u / %u mA",
3940 pdo_fixed_voltage(pdo), pdo_max, op, max);
3941 else
3942 tcpm_log(port,
3943 "Requested %u -> %u mV, %u mA for %u / %u mA",
3944 pdo_min_voltage(pdo), pdo_max_voltage(pdo),
3945 pdo_max, op, max);
3946 break;
3947 case PDO_TYPE_BATT:
3948 max = rdo_max_power(rdo);
3949 op = rdo_op_power(rdo);
3950 pdo_max = pdo_max_power(pdo);
3951
3952 if (op > pdo_max)
3953 return -EINVAL;
3954 if (max > pdo_max && !(rdo & RDO_CAP_MISMATCH))
3955 return -EINVAL;
3956 tcpm_log(port,
3957 "Requested %u -> %u mV, %u mW for %u / %u mW",
3958 pdo_min_voltage(pdo), pdo_max_voltage(pdo),
3959 pdo_max, op, max);
3960 break;
3961 default:
3962 return -EINVAL;
3963 }
3964
3965 port->op_vsafe5v = index == 1;
3966
3967 return 0;
3968 }
3969
3970 #define min_power(x, y) min(pdo_max_power(x), pdo_max_power(y))
3971 #define min_current(x, y) min(pdo_max_current(x), pdo_max_current(y))
3972
tcpm_pd_select_pdo(struct tcpm_port * port,int * sink_pdo,int * src_pdo)3973 static int tcpm_pd_select_pdo(struct tcpm_port *port, int *sink_pdo,
3974 int *src_pdo)
3975 {
3976 unsigned int i, j, max_src_mv = 0, min_src_mv = 0, max_mw = 0,
3977 max_mv = 0, src_mw = 0, src_ma = 0, max_snk_mv = 0,
3978 min_snk_mv = 0;
3979 int ret = -EINVAL;
3980
3981 port->pps_data.supported = false;
3982 port->usb_type = POWER_SUPPLY_USB_TYPE_PD;
3983 power_supply_changed(port->psy);
3984
3985 /*
3986 * Select the source PDO providing the most power which has a
3987 * matchig sink cap.
3988 */
3989 for (i = 0; i < port->nr_source_caps; i++) {
3990 u32 pdo = port->source_caps[i];
3991 enum pd_pdo_type type = pdo_type(pdo);
3992
3993 switch (type) {
3994 case PDO_TYPE_FIXED:
3995 max_src_mv = pdo_fixed_voltage(pdo);
3996 min_src_mv = max_src_mv;
3997 break;
3998 case PDO_TYPE_BATT:
3999 case PDO_TYPE_VAR:
4000 max_src_mv = pdo_max_voltage(pdo);
4001 min_src_mv = pdo_min_voltage(pdo);
4002 break;
4003 case PDO_TYPE_APDO:
4004 if (pdo_apdo_type(pdo) == APDO_TYPE_PPS) {
4005 port->pps_data.supported = true;
4006 port->usb_type =
4007 POWER_SUPPLY_USB_TYPE_PD_PPS;
4008 power_supply_changed(port->psy);
4009 }
4010 continue;
4011 default:
4012 tcpm_log(port, "Invalid source PDO type, ignoring");
4013 continue;
4014 }
4015
4016 switch (type) {
4017 case PDO_TYPE_FIXED:
4018 case PDO_TYPE_VAR:
4019 src_ma = pdo_max_current(pdo);
4020 src_mw = src_ma * min_src_mv / 1000;
4021 break;
4022 case PDO_TYPE_BATT:
4023 src_mw = pdo_max_power(pdo);
4024 break;
4025 case PDO_TYPE_APDO:
4026 continue;
4027 default:
4028 tcpm_log(port, "Invalid source PDO type, ignoring");
4029 continue;
4030 }
4031
4032 for (j = 0; j < port->nr_snk_pdo; j++) {
4033 pdo = port->snk_pdo[j];
4034
4035 switch (pdo_type(pdo)) {
4036 case PDO_TYPE_FIXED:
4037 max_snk_mv = pdo_fixed_voltage(pdo);
4038 min_snk_mv = max_snk_mv;
4039 break;
4040 case PDO_TYPE_BATT:
4041 case PDO_TYPE_VAR:
4042 max_snk_mv = pdo_max_voltage(pdo);
4043 min_snk_mv = pdo_min_voltage(pdo);
4044 break;
4045 case PDO_TYPE_APDO:
4046 continue;
4047 default:
4048 tcpm_log(port, "Invalid sink PDO type, ignoring");
4049 continue;
4050 }
4051
4052 if (max_src_mv <= max_snk_mv &&
4053 min_src_mv >= min_snk_mv) {
4054 /* Prefer higher voltages if available */
4055 if ((src_mw == max_mw && min_src_mv > max_mv) ||
4056 src_mw > max_mw) {
4057 *src_pdo = i;
4058 *sink_pdo = j;
4059 max_mw = src_mw;
4060 max_mv = min_src_mv;
4061 ret = 0;
4062 }
4063 }
4064 }
4065 }
4066
4067 return ret;
4068 }
4069
tcpm_pd_select_pps_apdo(struct tcpm_port * port)4070 static unsigned int tcpm_pd_select_pps_apdo(struct tcpm_port *port)
4071 {
4072 unsigned int i, src_ma, max_temp_mw = 0, max_op_ma, op_mw;
4073 unsigned int src_pdo = 0;
4074 u32 pdo, src;
4075
4076 for (i = 1; i < port->nr_source_caps; ++i) {
4077 pdo = port->source_caps[i];
4078
4079 switch (pdo_type(pdo)) {
4080 case PDO_TYPE_APDO:
4081 if (pdo_apdo_type(pdo) != APDO_TYPE_PPS) {
4082 tcpm_log(port, "Not PPS APDO (source), ignoring");
4083 continue;
4084 }
4085
4086 if (port->pps_data.req_out_volt > pdo_pps_apdo_max_voltage(pdo) ||
4087 port->pps_data.req_out_volt < pdo_pps_apdo_min_voltage(pdo))
4088 continue;
4089
4090 src_ma = pdo_pps_apdo_max_current(pdo);
4091 max_op_ma = min(src_ma, port->pps_data.req_op_curr);
4092 op_mw = max_op_ma * port->pps_data.req_out_volt / 1000;
4093 if (op_mw > max_temp_mw) {
4094 src_pdo = i;
4095 max_temp_mw = op_mw;
4096 }
4097 break;
4098 default:
4099 tcpm_log(port, "Not APDO type (source), ignoring");
4100 continue;
4101 }
4102 }
4103
4104 if (src_pdo) {
4105 src = port->source_caps[src_pdo];
4106
4107 port->pps_data.req_min_volt = pdo_pps_apdo_min_voltage(src);
4108 port->pps_data.req_max_volt = pdo_pps_apdo_max_voltage(src);
4109 port->pps_data.req_max_curr = pdo_pps_apdo_max_current(src);
4110 port->pps_data.req_op_curr = min(port->pps_data.req_max_curr,
4111 port->pps_data.req_op_curr);
4112 }
4113
4114 return src_pdo;
4115 }
4116
tcpm_pd_build_request(struct tcpm_port * port,u32 * rdo)4117 static int tcpm_pd_build_request(struct tcpm_port *port, u32 *rdo)
4118 {
4119 unsigned int mv, ma, mw, flags;
4120 unsigned int max_ma, max_mw;
4121 enum pd_pdo_type type;
4122 u32 pdo, matching_snk_pdo;
4123 int src_pdo_index = 0;
4124 int snk_pdo_index = 0;
4125 int ret;
4126
4127 ret = tcpm_pd_select_pdo(port, &snk_pdo_index, &src_pdo_index);
4128 if (ret < 0)
4129 return ret;
4130
4131 pdo = port->source_caps[src_pdo_index];
4132 matching_snk_pdo = port->snk_pdo[snk_pdo_index];
4133 type = pdo_type(pdo);
4134
4135 switch (type) {
4136 case PDO_TYPE_FIXED:
4137 mv = pdo_fixed_voltage(pdo);
4138 break;
4139 case PDO_TYPE_BATT:
4140 case PDO_TYPE_VAR:
4141 mv = pdo_min_voltage(pdo);
4142 break;
4143 default:
4144 tcpm_log(port, "Invalid PDO selected!");
4145 return -EINVAL;
4146 }
4147
4148 /* Select maximum available current within the sink pdo's limit */
4149 if (type == PDO_TYPE_BATT) {
4150 mw = min_power(pdo, matching_snk_pdo);
4151 ma = 1000 * mw / mv;
4152 } else {
4153 ma = min_current(pdo, matching_snk_pdo);
4154 mw = ma * mv / 1000;
4155 }
4156
4157 flags = RDO_USB_COMM | RDO_NO_SUSPEND;
4158
4159 /* Set mismatch bit if offered power is less than operating power */
4160 max_ma = ma;
4161 max_mw = mw;
4162 if (mw < port->operating_snk_mw) {
4163 flags |= RDO_CAP_MISMATCH;
4164 if (type == PDO_TYPE_BATT &&
4165 (pdo_max_power(matching_snk_pdo) > pdo_max_power(pdo)))
4166 max_mw = pdo_max_power(matching_snk_pdo);
4167 else if (pdo_max_current(matching_snk_pdo) >
4168 pdo_max_current(pdo))
4169 max_ma = pdo_max_current(matching_snk_pdo);
4170 }
4171
4172 tcpm_log(port, "cc=%d cc1=%d cc2=%d vbus=%d vconn=%s polarity=%d",
4173 port->cc_req, port->cc1, port->cc2, port->vbus_source,
4174 port->vconn_role == TYPEC_SOURCE ? "source" : "sink",
4175 port->polarity);
4176
4177 if (type == PDO_TYPE_BATT) {
4178 *rdo = RDO_BATT(src_pdo_index + 1, mw, max_mw, flags);
4179
4180 tcpm_log(port, "Requesting PDO %d: %u mV, %u mW%s",
4181 src_pdo_index, mv, mw,
4182 flags & RDO_CAP_MISMATCH ? " [mismatch]" : "");
4183 } else {
4184 *rdo = RDO_FIXED(src_pdo_index + 1, ma, max_ma, flags);
4185
4186 tcpm_log(port, "Requesting PDO %d: %u mV, %u mA%s",
4187 src_pdo_index, mv, ma,
4188 flags & RDO_CAP_MISMATCH ? " [mismatch]" : "");
4189 }
4190
4191 port->req_current_limit = ma;
4192 port->req_supply_voltage = mv;
4193
4194 return 0;
4195 }
4196
tcpm_pd_send_request(struct tcpm_port * port)4197 static int tcpm_pd_send_request(struct tcpm_port *port)
4198 {
4199 struct pd_message msg;
4200 int ret;
4201 u32 rdo;
4202
4203 ret = tcpm_pd_build_request(port, &rdo);
4204 if (ret < 0)
4205 return ret;
4206
4207 /*
4208 * Relax the threshold as voltage will be adjusted after Accept Message plus tSrcTransition.
4209 * It is safer to modify the threshold here.
4210 */
4211 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_USB, false, 0);
4212
4213 memset(&msg, 0, sizeof(msg));
4214 msg.header = PD_HEADER_LE(PD_DATA_REQUEST,
4215 port->pwr_role,
4216 port->data_role,
4217 port->negotiated_rev,
4218 port->message_id, 1);
4219 msg.payload[0] = cpu_to_le32(rdo);
4220
4221 return tcpm_pd_transmit(port, TCPC_TX_SOP, &msg);
4222 }
4223
tcpm_pd_build_pps_request(struct tcpm_port * port,u32 * rdo)4224 static int tcpm_pd_build_pps_request(struct tcpm_port *port, u32 *rdo)
4225 {
4226 unsigned int out_mv, op_ma, op_mw, max_mv, max_ma, flags;
4227 unsigned int src_pdo_index;
4228
4229 src_pdo_index = tcpm_pd_select_pps_apdo(port);
4230 if (!src_pdo_index)
4231 return -EOPNOTSUPP;
4232
4233 max_mv = port->pps_data.req_max_volt;
4234 max_ma = port->pps_data.req_max_curr;
4235 out_mv = port->pps_data.req_out_volt;
4236 op_ma = port->pps_data.req_op_curr;
4237
4238 flags = RDO_USB_COMM | RDO_NO_SUSPEND;
4239
4240 op_mw = (op_ma * out_mv) / 1000;
4241 if (op_mw < port->operating_snk_mw) {
4242 /*
4243 * Try raising current to meet power needs. If that's not enough
4244 * then try upping the voltage. If that's still not enough
4245 * then we've obviously chosen a PPS APDO which really isn't
4246 * suitable so abandon ship.
4247 */
4248 op_ma = (port->operating_snk_mw * 1000) / out_mv;
4249 if ((port->operating_snk_mw * 1000) % out_mv)
4250 ++op_ma;
4251 op_ma += RDO_PROG_CURR_MA_STEP - (op_ma % RDO_PROG_CURR_MA_STEP);
4252
4253 if (op_ma > max_ma) {
4254 op_ma = max_ma;
4255 out_mv = (port->operating_snk_mw * 1000) / op_ma;
4256 if ((port->operating_snk_mw * 1000) % op_ma)
4257 ++out_mv;
4258 out_mv += RDO_PROG_VOLT_MV_STEP -
4259 (out_mv % RDO_PROG_VOLT_MV_STEP);
4260
4261 if (out_mv > max_mv) {
4262 tcpm_log(port, "Invalid PPS APDO selected!");
4263 return -EINVAL;
4264 }
4265 }
4266 }
4267
4268 tcpm_log(port, "cc=%d cc1=%d cc2=%d vbus=%d vconn=%s polarity=%d",
4269 port->cc_req, port->cc1, port->cc2, port->vbus_source,
4270 port->vconn_role == TYPEC_SOURCE ? "source" : "sink",
4271 port->polarity);
4272
4273 *rdo = RDO_PROG(src_pdo_index + 1, out_mv, op_ma, flags);
4274
4275 tcpm_log(port, "Requesting APDO %d: %u mV, %u mA",
4276 src_pdo_index, out_mv, op_ma);
4277
4278 port->pps_data.req_op_curr = op_ma;
4279 port->pps_data.req_out_volt = out_mv;
4280
4281 return 0;
4282 }
4283
tcpm_pd_send_pps_request(struct tcpm_port * port)4284 static int tcpm_pd_send_pps_request(struct tcpm_port *port)
4285 {
4286 struct pd_message msg;
4287 int ret;
4288 u32 rdo;
4289
4290 ret = tcpm_pd_build_pps_request(port, &rdo);
4291 if (ret < 0)
4292 return ret;
4293
4294 /* Relax the threshold as voltage will be adjusted right after Accept Message. */
4295 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_USB, false, 0);
4296
4297 memset(&msg, 0, sizeof(msg));
4298 msg.header = PD_HEADER_LE(PD_DATA_REQUEST,
4299 port->pwr_role,
4300 port->data_role,
4301 port->negotiated_rev,
4302 port->message_id, 1);
4303 msg.payload[0] = cpu_to_le32(rdo);
4304
4305 return tcpm_pd_transmit(port, TCPC_TX_SOP, &msg);
4306 }
4307
tcpm_set_vbus(struct tcpm_port * port,bool enable)4308 static int tcpm_set_vbus(struct tcpm_port *port, bool enable)
4309 {
4310 int ret;
4311
4312 if (enable && port->vbus_charge)
4313 return -EINVAL;
4314
4315 tcpm_log(port, "vbus:=%d charge=%d", enable, port->vbus_charge);
4316
4317 ret = port->tcpc->set_vbus(port->tcpc, enable, port->vbus_charge);
4318 if (ret < 0)
4319 return ret;
4320
4321 port->vbus_source = enable;
4322 return 0;
4323 }
4324
tcpm_set_charge(struct tcpm_port * port,bool charge)4325 static int tcpm_set_charge(struct tcpm_port *port, bool charge)
4326 {
4327 int ret;
4328
4329 if (charge && port->vbus_source)
4330 return -EINVAL;
4331
4332 if (charge != port->vbus_charge) {
4333 tcpm_log(port, "vbus=%d charge:=%d", port->vbus_source, charge);
4334 ret = port->tcpc->set_vbus(port->tcpc, port->vbus_source,
4335 charge);
4336 if (ret < 0)
4337 return ret;
4338 }
4339 port->vbus_charge = charge;
4340 power_supply_changed(port->psy);
4341 return 0;
4342 }
4343
tcpm_start_toggling(struct tcpm_port * port,enum typec_cc_status cc)4344 static bool tcpm_start_toggling(struct tcpm_port *port, enum typec_cc_status cc)
4345 {
4346 int ret;
4347
4348 if (!port->tcpc->start_toggling)
4349 return false;
4350
4351 tcpm_log_force(port, "Start toggling");
4352 ret = port->tcpc->start_toggling(port->tcpc, port->port_type, cc);
4353 return ret == 0;
4354 }
4355
tcpm_init_vbus(struct tcpm_port * port)4356 static int tcpm_init_vbus(struct tcpm_port *port)
4357 {
4358 int ret;
4359
4360 ret = port->tcpc->set_vbus(port->tcpc, false, false);
4361 port->vbus_source = false;
4362 port->vbus_charge = false;
4363 return ret;
4364 }
4365
tcpm_init_vconn(struct tcpm_port * port)4366 static int tcpm_init_vconn(struct tcpm_port *port)
4367 {
4368 int ret;
4369
4370 ret = port->tcpc->set_vconn(port->tcpc, false);
4371 port->vconn_role = TYPEC_SINK;
4372 return ret;
4373 }
4374
tcpm_typec_connect(struct tcpm_port * port)4375 static void tcpm_typec_connect(struct tcpm_port *port)
4376 {
4377 struct typec_partner *partner;
4378
4379 if (!port->connected) {
4380 port->connected = true;
4381 /* Make sure we don't report stale identity information */
4382 memset(&port->partner_ident, 0, sizeof(port->partner_ident));
4383 port->partner_desc.usb_pd = port->pd_capable;
4384 if (tcpm_port_is_debug(port))
4385 port->partner_desc.accessory = TYPEC_ACCESSORY_DEBUG;
4386 else if (tcpm_port_is_audio(port))
4387 port->partner_desc.accessory = TYPEC_ACCESSORY_AUDIO;
4388 else
4389 port->partner_desc.accessory = TYPEC_ACCESSORY_NONE;
4390 partner = typec_register_partner(port->typec_port, &port->partner_desc);
4391 if (IS_ERR(partner)) {
4392 dev_err(port->dev, "Failed to register partner (%ld)\n", PTR_ERR(partner));
4393 return;
4394 }
4395
4396 port->partner = partner;
4397 typec_partner_set_usb_power_delivery(port->partner, port->partner_pd);
4398 }
4399 }
4400
tcpm_src_attach(struct tcpm_port * port)4401 static int tcpm_src_attach(struct tcpm_port *port)
4402 {
4403 enum typec_cc_polarity polarity =
4404 port->cc2 == TYPEC_CC_RD ? TYPEC_POLARITY_CC2
4405 : TYPEC_POLARITY_CC1;
4406 int ret;
4407
4408 if (port->attached)
4409 return 0;
4410
4411 ret = tcpm_set_polarity(port, polarity);
4412 if (ret < 0)
4413 return ret;
4414
4415 tcpm_enable_auto_vbus_discharge(port, true);
4416
4417 /*
4418 * USB Type-C specification, version 1.2,
4419 * chapter 4.5.2.2.8.1 (Attached.SRC Requirements)
4420 * Enable VCONN only if the non-RD port is set to RA.
4421 */
4422 if ((polarity == TYPEC_POLARITY_CC1 && port->cc2 == TYPEC_CC_RA) ||
4423 (polarity == TYPEC_POLARITY_CC2 && port->cc1 == TYPEC_CC_RA)) {
4424 ret = tcpm_set_vconn(port, true);
4425 if (ret < 0)
4426 return ret;
4427 }
4428
4429 ret = tcpm_set_vbus(port, true);
4430 if (ret < 0)
4431 goto out_disable_vconn;
4432
4433 ret = tcpm_set_roles(port, true, TYPEC_STATE_USB, TYPEC_SOURCE,
4434 tcpm_data_role_for_source(port));
4435 if (ret < 0)
4436 goto out_disable_vbus;
4437
4438 if (port->pd_supported) {
4439 ret = port->tcpc->set_pd_rx(port->tcpc, true);
4440 if (ret < 0)
4441 goto out_disable_mux;
4442 }
4443
4444 port->pd_capable = false;
4445
4446 port->partner = NULL;
4447
4448 port->attached = true;
4449 port->send_discover = true;
4450 port->send_discover_prime = false;
4451
4452 return 0;
4453
4454 out_disable_mux:
4455 tcpm_mux_set(port, TYPEC_STATE_SAFE, USB_ROLE_NONE,
4456 TYPEC_ORIENTATION_NONE);
4457 out_disable_vbus:
4458 tcpm_set_vbus(port, false);
4459 out_disable_vconn:
4460 tcpm_set_vconn(port, false);
4461
4462 return ret;
4463 }
4464
tcpm_typec_disconnect(struct tcpm_port * port)4465 static void tcpm_typec_disconnect(struct tcpm_port *port)
4466 {
4467 /*
4468 * Unregister plug/cable outside of port->connected because cable can
4469 * be discovered before SRC_READY/SNK_READY states where port->connected
4470 * is set.
4471 */
4472 typec_unregister_plug(port->plug_prime);
4473 typec_unregister_cable(port->cable);
4474 port->plug_prime = NULL;
4475 port->cable = NULL;
4476 if (port->connected) {
4477 if (port->partner) {
4478 typec_partner_set_usb_power_delivery(port->partner, NULL);
4479 typec_unregister_partner(port->partner);
4480 port->partner = NULL;
4481 }
4482 port->connected = false;
4483 }
4484 }
4485
tcpm_unregister_altmodes(struct tcpm_port * port)4486 static void tcpm_unregister_altmodes(struct tcpm_port *port)
4487 {
4488 struct pd_mode_data *modep = &port->mode_data;
4489 struct pd_mode_data *modep_prime = &port->mode_data_prime;
4490 int i;
4491
4492 for (i = 0; i < modep->altmodes; i++) {
4493 typec_unregister_altmode(port->partner_altmode[i]);
4494 port->partner_altmode[i] = NULL;
4495 }
4496 for (i = 0; i < modep_prime->altmodes; i++) {
4497 typec_unregister_altmode(port->plug_prime_altmode[i]);
4498 port->plug_prime_altmode[i] = NULL;
4499 }
4500
4501 memset(modep, 0, sizeof(*modep));
4502 memset(modep_prime, 0, sizeof(*modep_prime));
4503 }
4504
tcpm_set_partner_usb_comm_capable(struct tcpm_port * port,bool capable)4505 static void tcpm_set_partner_usb_comm_capable(struct tcpm_port *port, bool capable)
4506 {
4507 tcpm_log(port, "Setting usb_comm capable %s", str_true_false(capable));
4508
4509 if (port->tcpc->set_partner_usb_comm_capable)
4510 port->tcpc->set_partner_usb_comm_capable(port->tcpc, capable);
4511 }
4512
tcpm_reset_port(struct tcpm_port * port)4513 static void tcpm_reset_port(struct tcpm_port *port)
4514 {
4515 tcpm_enable_auto_vbus_discharge(port, false);
4516 port->in_ams = false;
4517 port->ams = NONE_AMS;
4518 port->vdm_sm_running = false;
4519 tcpm_unregister_altmodes(port);
4520 tcpm_typec_disconnect(port);
4521 port->attached = false;
4522 port->pd_capable = false;
4523 port->pps_data.supported = false;
4524 tcpm_set_partner_usb_comm_capable(port, false);
4525
4526 /*
4527 * First Rx ID should be 0; set this to a sentinel of -1 so that
4528 * we can check tcpm_pd_rx_handler() if we had seen it before.
4529 */
4530 port->rx_msgid = -1;
4531 port->rx_msgid_prime = -1;
4532
4533 port->tcpc->set_pd_rx(port->tcpc, false);
4534 tcpm_init_vbus(port); /* also disables charging */
4535 tcpm_init_vconn(port);
4536 tcpm_set_current_limit(port, 0, 0);
4537 tcpm_set_polarity(port, TYPEC_POLARITY_CC1);
4538 tcpm_mux_set(port, TYPEC_STATE_SAFE, USB_ROLE_NONE,
4539 TYPEC_ORIENTATION_NONE);
4540 tcpm_set_attached_state(port, false);
4541 port->try_src_count = 0;
4542 port->try_snk_count = 0;
4543 port->usb_type = POWER_SUPPLY_USB_TYPE_C;
4544 power_supply_changed(port->psy);
4545 port->nr_sink_caps = 0;
4546 port->sink_cap_done = false;
4547 if (port->tcpc->enable_frs)
4548 port->tcpc->enable_frs(port->tcpc, false);
4549
4550 usb_power_delivery_unregister_capabilities(port->partner_sink_caps);
4551 port->partner_sink_caps = NULL;
4552 usb_power_delivery_unregister_capabilities(port->partner_source_caps);
4553 port->partner_source_caps = NULL;
4554 usb_power_delivery_unregister(port->partner_pd);
4555 port->partner_pd = NULL;
4556 }
4557
tcpm_detach(struct tcpm_port * port)4558 static void tcpm_detach(struct tcpm_port *port)
4559 {
4560 if (tcpm_port_is_disconnected(port))
4561 port->hard_reset_count = 0;
4562
4563 if (!port->attached)
4564 return;
4565
4566 if (port->tcpc->set_bist_data) {
4567 tcpm_log(port, "disable BIST MODE TESTDATA");
4568 port->tcpc->set_bist_data(port->tcpc, false);
4569 }
4570
4571 tcpm_reset_port(port);
4572 }
4573
tcpm_src_detach(struct tcpm_port * port)4574 static void tcpm_src_detach(struct tcpm_port *port)
4575 {
4576 tcpm_detach(port);
4577 }
4578
tcpm_snk_attach(struct tcpm_port * port)4579 static int tcpm_snk_attach(struct tcpm_port *port)
4580 {
4581 int ret;
4582
4583 if (port->attached)
4584 return 0;
4585
4586 ret = tcpm_set_polarity(port, port->cc2 != TYPEC_CC_OPEN ?
4587 TYPEC_POLARITY_CC2 : TYPEC_POLARITY_CC1);
4588 if (ret < 0)
4589 return ret;
4590
4591 tcpm_enable_auto_vbus_discharge(port, true);
4592
4593 ret = tcpm_set_roles(port, true, TYPEC_STATE_USB,
4594 TYPEC_SINK, tcpm_data_role_for_sink(port));
4595 if (ret < 0)
4596 return ret;
4597
4598 port->pd_capable = false;
4599
4600 port->partner = NULL;
4601
4602 port->attached = true;
4603 port->send_discover = true;
4604 port->send_discover_prime = false;
4605
4606 return 0;
4607 }
4608
tcpm_snk_detach(struct tcpm_port * port)4609 static void tcpm_snk_detach(struct tcpm_port *port)
4610 {
4611 tcpm_detach(port);
4612 }
4613
tcpm_acc_attach(struct tcpm_port * port)4614 static int tcpm_acc_attach(struct tcpm_port *port)
4615 {
4616 int ret;
4617 enum typec_role role;
4618 enum typec_data_role data;
4619 int state = TYPEC_STATE_USB;
4620
4621 if (port->attached)
4622 return 0;
4623
4624 role = tcpm_port_is_sink(port) ? TYPEC_SINK : TYPEC_SOURCE;
4625 data = tcpm_port_is_sink(port) ? tcpm_data_role_for_sink(port)
4626 : tcpm_data_role_for_source(port);
4627
4628 if (tcpm_port_is_audio(port))
4629 state = TYPEC_MODE_AUDIO;
4630
4631 if (tcpm_port_is_debug(port))
4632 state = TYPEC_MODE_DEBUG;
4633
4634 ret = tcpm_set_roles(port, true, state, role, data);
4635 if (ret < 0)
4636 return ret;
4637
4638 port->partner = NULL;
4639
4640 tcpm_typec_connect(port);
4641
4642 port->attached = true;
4643
4644 return 0;
4645 }
4646
tcpm_acc_detach(struct tcpm_port * port)4647 static void tcpm_acc_detach(struct tcpm_port *port)
4648 {
4649 tcpm_detach(port);
4650 }
4651
hard_reset_state(struct tcpm_port * port)4652 static inline enum tcpm_state hard_reset_state(struct tcpm_port *port)
4653 {
4654 if (port->hard_reset_count < PD_N_HARD_RESET_COUNT)
4655 return HARD_RESET_SEND;
4656 if (port->pd_capable)
4657 return ERROR_RECOVERY;
4658 if (port->pwr_role == TYPEC_SOURCE)
4659 return SRC_UNATTACHED;
4660 if (port->state == SNK_WAIT_CAPABILITIES ||
4661 port->state == SNK_WAIT_CAPABILITIES_TIMEOUT)
4662 return SNK_READY;
4663 return SNK_UNATTACHED;
4664 }
4665
unattached_state(struct tcpm_port * port)4666 static inline enum tcpm_state unattached_state(struct tcpm_port *port)
4667 {
4668 if (port->port_type == TYPEC_PORT_DRP) {
4669 if (port->pwr_role == TYPEC_SOURCE)
4670 return SRC_UNATTACHED;
4671 else
4672 return SNK_UNATTACHED;
4673 } else if (port->port_type == TYPEC_PORT_SRC) {
4674 return SRC_UNATTACHED;
4675 }
4676
4677 return SNK_UNATTACHED;
4678 }
4679
tcpm_swap_complete(struct tcpm_port * port,int result)4680 static void tcpm_swap_complete(struct tcpm_port *port, int result)
4681 {
4682 if (port->swap_pending) {
4683 port->swap_status = result;
4684 port->swap_pending = false;
4685 port->non_pd_role_swap = false;
4686 complete(&port->swap_complete);
4687 }
4688 }
4689
tcpm_get_pwr_opmode(enum typec_cc_status cc)4690 static enum typec_pwr_opmode tcpm_get_pwr_opmode(enum typec_cc_status cc)
4691 {
4692 switch (cc) {
4693 case TYPEC_CC_RP_1_5:
4694 return TYPEC_PWR_MODE_1_5A;
4695 case TYPEC_CC_RP_3_0:
4696 return TYPEC_PWR_MODE_3_0A;
4697 case TYPEC_CC_RP_DEF:
4698 default:
4699 return TYPEC_PWR_MODE_USB;
4700 }
4701 }
4702
tcpm_pwr_opmode_to_rp(enum typec_pwr_opmode opmode)4703 static enum typec_cc_status tcpm_pwr_opmode_to_rp(enum typec_pwr_opmode opmode)
4704 {
4705 switch (opmode) {
4706 case TYPEC_PWR_MODE_USB:
4707 return TYPEC_CC_RP_DEF;
4708 case TYPEC_PWR_MODE_1_5A:
4709 return TYPEC_CC_RP_1_5;
4710 case TYPEC_PWR_MODE_3_0A:
4711 case TYPEC_PWR_MODE_PD:
4712 default:
4713 return TYPEC_CC_RP_3_0;
4714 }
4715 }
4716
tcpm_set_initial_svdm_version(struct tcpm_port * port)4717 static void tcpm_set_initial_svdm_version(struct tcpm_port *port)
4718 {
4719 if (!port->partner)
4720 return;
4721
4722 switch (port->negotiated_rev) {
4723 case PD_REV30:
4724 break;
4725 /*
4726 * 6.4.4.2.3 Structured VDM Version
4727 * 2.0 states "At this time, there is only one version (1.0) defined.
4728 * This field Shall be set to zero to indicate Version 1.0."
4729 * 3.0 states "This field Shall be set to 01b to indicate Version 2.0."
4730 * To ensure that we follow the Power Delivery revision we are currently
4731 * operating on, downgrade the SVDM version to the highest one supported
4732 * by the Power Delivery revision.
4733 */
4734 case PD_REV20:
4735 typec_partner_set_svdm_version(port->partner, SVDM_VER_1_0);
4736 break;
4737 default:
4738 typec_partner_set_svdm_version(port->partner, SVDM_VER_1_0);
4739 break;
4740 }
4741 }
4742
tcpm_set_initial_negotiated_rev(struct tcpm_port * port)4743 static void tcpm_set_initial_negotiated_rev(struct tcpm_port *port)
4744 {
4745 switch (port->pd_rev.rev_major) {
4746 case PD_CAP_REV10:
4747 port->negotiated_rev = PD_REV10;
4748 break;
4749 case PD_CAP_REV20:
4750 port->negotiated_rev = PD_REV20;
4751 break;
4752 case PD_CAP_REV30:
4753 port->negotiated_rev = PD_REV30;
4754 break;
4755 default:
4756 port->negotiated_rev = PD_MAX_REV;
4757 break;
4758 }
4759 port->negotiated_rev_prime = port->negotiated_rev;
4760 }
4761
run_state_machine(struct tcpm_port * port)4762 static void run_state_machine(struct tcpm_port *port)
4763 {
4764 int ret;
4765 enum typec_pwr_opmode opmode;
4766 unsigned int msecs;
4767 enum tcpm_state upcoming_state;
4768
4769 if (port->tcpc->check_contaminant && port->state != CHECK_CONTAMINANT)
4770 port->potential_contaminant = ((port->enter_state == SRC_ATTACH_WAIT &&
4771 port->state == SRC_UNATTACHED) ||
4772 (port->enter_state == SNK_ATTACH_WAIT &&
4773 port->state == SNK_UNATTACHED) ||
4774 (port->enter_state == SNK_DEBOUNCED &&
4775 port->state == SNK_UNATTACHED));
4776
4777 port->enter_state = port->state;
4778 switch (port->state) {
4779 case TOGGLING:
4780 break;
4781 case CHECK_CONTAMINANT:
4782 port->tcpc->check_contaminant(port->tcpc);
4783 break;
4784 /* SRC states */
4785 case SRC_UNATTACHED:
4786 if (!port->non_pd_role_swap)
4787 tcpm_swap_complete(port, -ENOTCONN);
4788 tcpm_src_detach(port);
4789 if (port->potential_contaminant) {
4790 tcpm_set_state(port, CHECK_CONTAMINANT, 0);
4791 break;
4792 }
4793 if (tcpm_start_toggling(port, tcpm_rp_cc(port))) {
4794 tcpm_set_state(port, TOGGLING, 0);
4795 break;
4796 }
4797 tcpm_set_cc(port, tcpm_rp_cc(port));
4798 if (port->port_type == TYPEC_PORT_DRP)
4799 tcpm_set_state(port, SNK_UNATTACHED, PD_T_DRP_SNK);
4800 break;
4801 case SRC_ATTACH_WAIT:
4802 if (tcpm_port_is_debug(port))
4803 tcpm_set_state(port, DEBUG_ACC_ATTACHED,
4804 port->timings.cc_debounce_time);
4805 else if (tcpm_port_is_audio(port))
4806 tcpm_set_state(port, AUDIO_ACC_ATTACHED,
4807 port->timings.cc_debounce_time);
4808 else if (tcpm_port_is_source(port) && port->vbus_vsafe0v)
4809 tcpm_set_state(port,
4810 tcpm_try_snk(port) ? SNK_TRY
4811 : SRC_ATTACHED,
4812 port->timings.cc_debounce_time);
4813 break;
4814
4815 case SNK_TRY:
4816 port->try_snk_count++;
4817 /*
4818 * Requirements:
4819 * - Do not drive vconn or vbus
4820 * - Terminate CC pins (both) to Rd
4821 * Action:
4822 * - Wait for tDRPTry (PD_T_DRP_TRY).
4823 * Until then, ignore any state changes.
4824 */
4825 tcpm_set_cc(port, TYPEC_CC_RD);
4826 tcpm_set_state(port, SNK_TRY_WAIT, PD_T_DRP_TRY);
4827 break;
4828 case SNK_TRY_WAIT:
4829 if (tcpm_port_is_sink(port)) {
4830 tcpm_set_state(port, SNK_TRY_WAIT_DEBOUNCE, 0);
4831 } else {
4832 tcpm_set_state(port, SRC_TRYWAIT, 0);
4833 port->max_wait = 0;
4834 }
4835 break;
4836 case SNK_TRY_WAIT_DEBOUNCE:
4837 tcpm_set_state(port, SNK_TRY_WAIT_DEBOUNCE_CHECK_VBUS,
4838 PD_T_TRY_CC_DEBOUNCE);
4839 break;
4840 case SNK_TRY_WAIT_DEBOUNCE_CHECK_VBUS:
4841 if (port->vbus_present && tcpm_port_is_sink(port))
4842 tcpm_set_state(port, SNK_ATTACHED, 0);
4843 else
4844 port->max_wait = 0;
4845 break;
4846 case SRC_TRYWAIT:
4847 tcpm_set_cc(port, tcpm_rp_cc(port));
4848 if (port->max_wait == 0) {
4849 port->max_wait = jiffies +
4850 msecs_to_jiffies(PD_T_DRP_TRY);
4851 tcpm_set_state(port, SRC_TRYWAIT_UNATTACHED,
4852 PD_T_DRP_TRY);
4853 } else {
4854 if (time_is_after_jiffies(port->max_wait))
4855 tcpm_set_state(port, SRC_TRYWAIT_UNATTACHED,
4856 jiffies_to_msecs(port->max_wait -
4857 jiffies));
4858 else
4859 tcpm_set_state(port, SNK_UNATTACHED, 0);
4860 }
4861 break;
4862 case SRC_TRYWAIT_DEBOUNCE:
4863 tcpm_set_state(port, SRC_ATTACHED, port->timings.cc_debounce_time);
4864 break;
4865 case SRC_TRYWAIT_UNATTACHED:
4866 tcpm_set_state(port, SNK_UNATTACHED, 0);
4867 break;
4868
4869 case SRC_ATTACHED:
4870 ret = tcpm_src_attach(port);
4871 tcpm_set_state(port, SRC_UNATTACHED,
4872 ret < 0 ? 0 : PD_T_PS_SOURCE_ON);
4873 break;
4874 case SRC_STARTUP:
4875 opmode = tcpm_get_pwr_opmode(tcpm_rp_cc(port));
4876 typec_set_pwr_opmode(port->typec_port, opmode);
4877 port->pwr_opmode = TYPEC_PWR_MODE_USB;
4878 port->caps_count = 0;
4879 tcpm_set_initial_negotiated_rev(port);
4880 port->message_id = 0;
4881 port->message_id_prime = 0;
4882 port->rx_msgid = -1;
4883 port->rx_msgid_prime = -1;
4884 port->explicit_contract = false;
4885 /* SNK -> SRC POWER/FAST_ROLE_SWAP finished */
4886 if (port->ams == POWER_ROLE_SWAP ||
4887 port->ams == FAST_ROLE_SWAP)
4888 tcpm_ams_finish(port);
4889 if (!port->pd_supported) {
4890 tcpm_set_state(port, SRC_READY, 0);
4891 break;
4892 }
4893 port->upcoming_state = SRC_SEND_CAPABILITIES;
4894 tcpm_ams_start(port, POWER_NEGOTIATION);
4895 break;
4896 case SRC_SEND_CAPABILITIES:
4897 port->caps_count++;
4898 if (port->caps_count > PD_N_CAPS_COUNT) {
4899 tcpm_set_state(port, SRC_READY, 0);
4900 break;
4901 }
4902 ret = tcpm_pd_send_source_caps(port);
4903 if (ret < 0) {
4904 if (tcpm_can_communicate_sop_prime(port) &&
4905 IS_ERR_OR_NULL(port->cable))
4906 tcpm_set_state(port, SRC_VDM_IDENTITY_REQUEST, 0);
4907 else
4908 tcpm_set_state(port, SRC_SEND_CAPABILITIES,
4909 PD_T_SEND_SOURCE_CAP);
4910 } else {
4911 /*
4912 * Per standard, we should clear the reset counter here.
4913 * However, that can result in state machine hang-ups.
4914 * Reset it only in READY state to improve stability.
4915 */
4916 /* port->hard_reset_count = 0; */
4917 port->caps_count = 0;
4918 port->pd_capable = true;
4919 tcpm_set_state_cond(port, SRC_SEND_CAPABILITIES_TIMEOUT,
4920 PD_T_SENDER_RESPONSE);
4921 }
4922 break;
4923 case SRC_SEND_CAPABILITIES_TIMEOUT:
4924 /*
4925 * Error recovery for a PD_DATA_SOURCE_CAP reply timeout.
4926 *
4927 * PD 2.0 sinks are supposed to accept src-capabilities with a
4928 * 3.0 header and simply ignore any src PDOs which the sink does
4929 * not understand such as PPS but some 2.0 sinks instead ignore
4930 * the entire PD_DATA_SOURCE_CAP message, causing contract
4931 * negotiation to fail.
4932 *
4933 * After PD_N_HARD_RESET_COUNT hard-reset attempts, we try
4934 * sending src-capabilities with a lower PD revision to
4935 * make these broken sinks work.
4936 */
4937 if (port->hard_reset_count < PD_N_HARD_RESET_COUNT) {
4938 tcpm_set_state(port, HARD_RESET_SEND, 0);
4939 } else if (port->negotiated_rev > PD_REV20) {
4940 port->negotiated_rev--;
4941 port->hard_reset_count = 0;
4942 tcpm_set_state(port, SRC_SEND_CAPABILITIES, 0);
4943 } else {
4944 tcpm_set_state(port, hard_reset_state(port), 0);
4945 }
4946 break;
4947 case SRC_NEGOTIATE_CAPABILITIES:
4948 ret = tcpm_pd_check_request(port);
4949 if (ret < 0) {
4950 tcpm_pd_send_control(port, PD_CTRL_REJECT, TCPC_TX_SOP);
4951 if (!port->explicit_contract) {
4952 tcpm_set_state(port,
4953 SRC_WAIT_NEW_CAPABILITIES, 0);
4954 } else {
4955 tcpm_set_state(port, SRC_READY, 0);
4956 }
4957 } else {
4958 tcpm_pd_send_control(port, PD_CTRL_ACCEPT, TCPC_TX_SOP);
4959 tcpm_set_partner_usb_comm_capable(port,
4960 !!(port->sink_request & RDO_USB_COMM));
4961 tcpm_set_state(port, SRC_TRANSITION_SUPPLY,
4962 PD_T_SRC_TRANSITION);
4963 }
4964 break;
4965 case SRC_TRANSITION_SUPPLY:
4966 /* XXX: regulator_set_voltage(vbus, ...) */
4967 tcpm_pd_send_control(port, PD_CTRL_PS_RDY, TCPC_TX_SOP);
4968 port->explicit_contract = true;
4969 typec_set_pwr_opmode(port->typec_port, TYPEC_PWR_MODE_PD);
4970 port->pwr_opmode = TYPEC_PWR_MODE_PD;
4971 tcpm_set_state_cond(port, SRC_READY, 0);
4972 break;
4973 case SRC_READY:
4974 #if 1
4975 port->hard_reset_count = 0;
4976 #endif
4977 port->try_src_count = 0;
4978
4979 tcpm_swap_complete(port, 0);
4980 tcpm_typec_connect(port);
4981
4982 if (port->ams != NONE_AMS)
4983 tcpm_ams_finish(port);
4984 if (port->next_ams != NONE_AMS) {
4985 port->ams = port->next_ams;
4986 port->next_ams = NONE_AMS;
4987 }
4988
4989 /*
4990 * If previous AMS is interrupted, switch to the upcoming
4991 * state.
4992 */
4993 if (port->upcoming_state != INVALID_STATE) {
4994 upcoming_state = port->upcoming_state;
4995 port->upcoming_state = INVALID_STATE;
4996 tcpm_set_state(port, upcoming_state, 0);
4997 break;
4998 }
4999
5000 /*
5001 * 6.4.4.3.1 Discover Identity
5002 * "The Discover Identity Command Shall only be sent to SOP when there is an
5003 * Explicit Contract."
5004 *
5005 * Discover Identity on SOP' should be discovered prior to the
5006 * ready state, but if done after a Vconn Swap following Discover
5007 * Identity on SOP then the discovery process can be run here
5008 * as well.
5009 */
5010 if (port->explicit_contract) {
5011 if (port->send_discover_prime) {
5012 port->tx_sop_type = TCPC_TX_SOP_PRIME;
5013 } else {
5014 port->tx_sop_type = TCPC_TX_SOP;
5015 tcpm_set_initial_svdm_version(port);
5016 }
5017 mod_send_discover_delayed_work(port, 0);
5018 } else {
5019 port->send_discover = false;
5020 port->send_discover_prime = false;
5021 }
5022
5023 /*
5024 * 6.3.5
5025 * Sending ping messages is not necessary if
5026 * - the source operates at vSafe5V
5027 * or
5028 * - The system is not operating in PD mode
5029 * or
5030 * - Both partners are connected using a Type-C connector
5031 *
5032 * There is no actual need to send PD messages since the local
5033 * port type-c and the spec does not clearly say whether PD is
5034 * possible when type-c is connected to Type-A/B
5035 */
5036 break;
5037 case SRC_WAIT_NEW_CAPABILITIES:
5038 /* Nothing to do... */
5039 break;
5040
5041 /* SNK states */
5042 case SNK_UNATTACHED:
5043 if (!port->non_pd_role_swap)
5044 tcpm_swap_complete(port, -ENOTCONN);
5045 tcpm_pps_complete(port, -ENOTCONN);
5046 tcpm_snk_detach(port);
5047 if (port->potential_contaminant) {
5048 tcpm_set_state(port, CHECK_CONTAMINANT, 0);
5049 break;
5050 }
5051 if (tcpm_start_toggling(port, TYPEC_CC_RD)) {
5052 tcpm_set_state(port, TOGGLING, 0);
5053 break;
5054 }
5055 tcpm_set_cc(port, TYPEC_CC_RD);
5056 if (port->port_type == TYPEC_PORT_DRP)
5057 tcpm_set_state(port, SRC_UNATTACHED, PD_T_DRP_SRC);
5058 break;
5059 case SNK_ATTACH_WAIT:
5060 if (tcpm_port_is_debug(port))
5061 tcpm_set_state(port, DEBUG_ACC_ATTACHED,
5062 PD_T_CC_DEBOUNCE);
5063 else if (tcpm_port_is_audio(port))
5064 tcpm_set_state(port, AUDIO_ACC_ATTACHED,
5065 PD_T_CC_DEBOUNCE);
5066 else if ((port->cc1 == TYPEC_CC_OPEN &&
5067 port->cc2 != TYPEC_CC_OPEN) ||
5068 (port->cc1 != TYPEC_CC_OPEN &&
5069 port->cc2 == TYPEC_CC_OPEN))
5070 tcpm_set_state(port, SNK_DEBOUNCED,
5071 port->timings.cc_debounce_time);
5072 else if (tcpm_port_is_disconnected(port))
5073 tcpm_set_state(port, SNK_UNATTACHED,
5074 PD_T_PD_DEBOUNCE);
5075 break;
5076 case SNK_DEBOUNCED:
5077 if (tcpm_port_is_disconnected(port))
5078 tcpm_set_state(port, SNK_UNATTACHED,
5079 PD_T_PD_DEBOUNCE);
5080 else if (tcpm_port_is_debug(port))
5081 tcpm_set_state(port, DEBUG_ACC_ATTACHED,
5082 PD_T_CC_DEBOUNCE);
5083 else if (tcpm_port_is_audio(port))
5084 tcpm_set_state(port, AUDIO_ACC_ATTACHED,
5085 PD_T_CC_DEBOUNCE);
5086 else if (port->vbus_present)
5087 tcpm_set_state(port,
5088 tcpm_try_src(port) ? SRC_TRY
5089 : SNK_ATTACHED,
5090 0);
5091 break;
5092 case SRC_TRY:
5093 port->try_src_count++;
5094 tcpm_set_cc(port, tcpm_rp_cc(port));
5095 port->max_wait = 0;
5096 tcpm_set_state(port, SRC_TRY_WAIT, 0);
5097 break;
5098 case SRC_TRY_WAIT:
5099 if (port->max_wait == 0) {
5100 port->max_wait = jiffies +
5101 msecs_to_jiffies(PD_T_DRP_TRY);
5102 msecs = PD_T_DRP_TRY;
5103 } else {
5104 if (time_is_after_jiffies(port->max_wait))
5105 msecs = jiffies_to_msecs(port->max_wait -
5106 jiffies);
5107 else
5108 msecs = 0;
5109 }
5110 tcpm_set_state(port, SNK_TRYWAIT, msecs);
5111 break;
5112 case SRC_TRY_DEBOUNCE:
5113 tcpm_set_state(port, SRC_ATTACHED, PD_T_PD_DEBOUNCE);
5114 break;
5115 case SNK_TRYWAIT:
5116 tcpm_set_cc(port, TYPEC_CC_RD);
5117 tcpm_set_state(port, SNK_TRYWAIT_VBUS, port->timings.cc_debounce_time);
5118 break;
5119 case SNK_TRYWAIT_VBUS:
5120 /*
5121 * TCPM stays in this state indefinitely until VBUS
5122 * is detected as long as Rp is not detected for
5123 * more than a time period of tPDDebounce.
5124 */
5125 if (port->vbus_present && tcpm_port_is_sink(port)) {
5126 tcpm_set_state(port, SNK_ATTACHED, 0);
5127 break;
5128 }
5129 if (!tcpm_port_is_sink(port))
5130 tcpm_set_state(port, SNK_TRYWAIT_DEBOUNCE, 0);
5131 break;
5132 case SNK_TRYWAIT_DEBOUNCE:
5133 tcpm_set_state(port, SNK_UNATTACHED, PD_T_PD_DEBOUNCE);
5134 break;
5135 case SNK_ATTACHED:
5136 ret = tcpm_snk_attach(port);
5137 if (ret < 0)
5138 tcpm_set_state(port, SNK_UNATTACHED, 0);
5139 else
5140 /*
5141 * For Type C port controllers that use Battery Charging
5142 * Detection (based on BCv1.2 spec) to detect USB
5143 * charger type, add a delay of "snk_bc12_cmpletion_time"
5144 * before transitioning to SNK_STARTUP to allow BC1.2
5145 * detection to complete before PD is eventually enabled
5146 * in later states.
5147 */
5148 tcpm_set_state(port, SNK_STARTUP,
5149 port->timings.snk_bc12_cmpletion_time);
5150 break;
5151 case SNK_STARTUP:
5152 opmode = tcpm_get_pwr_opmode(port->polarity ?
5153 port->cc2 : port->cc1);
5154 typec_set_pwr_opmode(port->typec_port, opmode);
5155 port->pwr_opmode = TYPEC_PWR_MODE_USB;
5156 tcpm_set_initial_negotiated_rev(port);
5157 port->message_id = 0;
5158 port->message_id_prime = 0;
5159 port->rx_msgid = -1;
5160 port->rx_msgid_prime = -1;
5161 port->explicit_contract = false;
5162
5163 if (port->ams == POWER_ROLE_SWAP ||
5164 port->ams == FAST_ROLE_SWAP)
5165 /* SRC -> SNK POWER/FAST_ROLE_SWAP finished */
5166 tcpm_ams_finish(port);
5167
5168 tcpm_set_state(port, SNK_DISCOVERY, 0);
5169 break;
5170 case SNK_DISCOVERY:
5171 if (port->vbus_present) {
5172 u32 current_lim = tcpm_get_current_limit(port);
5173
5174 if (port->slow_charger_loop && (current_lim > PD_P_SNK_STDBY_MW / 5))
5175 current_lim = PD_P_SNK_STDBY_MW / 5;
5176 tcpm_set_current_limit(port, current_lim, 5000);
5177 /* Not sink vbus if operational current is 0mA */
5178 tcpm_set_charge(port, !port->pd_supported ||
5179 pdo_max_current(port->snk_pdo[0]));
5180
5181 if (!port->pd_supported)
5182 tcpm_set_state(port, SNK_READY, 0);
5183 else
5184 tcpm_set_state(port, SNK_WAIT_CAPABILITIES, 0);
5185 break;
5186 }
5187 /*
5188 * For DRP, timeouts differ. Also, handling is supposed to be
5189 * different and much more complex (dead battery detection;
5190 * see USB power delivery specification, section 8.3.3.6.1.5.1).
5191 */
5192 tcpm_set_state(port, hard_reset_state(port),
5193 port->port_type == TYPEC_PORT_DRP ?
5194 PD_T_DB_DETECT : PD_T_NO_RESPONSE);
5195 break;
5196 case SNK_DISCOVERY_DEBOUNCE:
5197 tcpm_set_state(port, SNK_DISCOVERY_DEBOUNCE_DONE,
5198 port->timings.cc_debounce_time);
5199 break;
5200 case SNK_DISCOVERY_DEBOUNCE_DONE:
5201 if (!tcpm_port_is_disconnected(port) &&
5202 tcpm_port_is_sink(port) &&
5203 ktime_after(port->delayed_runtime, ktime_get())) {
5204 tcpm_set_state(port, SNK_DISCOVERY,
5205 ktime_to_ms(ktime_sub(port->delayed_runtime, ktime_get())));
5206 break;
5207 }
5208 tcpm_set_state(port, unattached_state(port), 0);
5209 break;
5210 case SNK_WAIT_CAPABILITIES:
5211 ret = port->tcpc->set_pd_rx(port->tcpc, true);
5212 if (ret < 0) {
5213 tcpm_set_state(port, SNK_READY, 0);
5214 break;
5215 }
5216 /*
5217 * If VBUS has never been low, and we time out waiting
5218 * for source cap, try a soft reset first, in case we
5219 * were already in a stable contract before this boot.
5220 * Do this only once.
5221 */
5222 if (port->vbus_never_low) {
5223 port->vbus_never_low = false;
5224 upcoming_state = SNK_SOFT_RESET;
5225 } else {
5226 if (!port->self_powered)
5227 upcoming_state = SNK_WAIT_CAPABILITIES_TIMEOUT;
5228 else
5229 upcoming_state = hard_reset_state(port);
5230 }
5231
5232 tcpm_set_state(port, upcoming_state,
5233 port->timings.sink_wait_cap_time);
5234 break;
5235 case SNK_WAIT_CAPABILITIES_TIMEOUT:
5236 /*
5237 * There are some USB PD sources in the field, which do not
5238 * properly implement the specification and fail to start
5239 * sending Source Capability messages after a soft reset. The
5240 * specification suggests to do a hard reset when no Source
5241 * capability message is received within PD_T_SINK_WAIT_CAP,
5242 * but that might effectively kil the machine's power source.
5243 *
5244 * This slightly diverges from the specification and tries to
5245 * recover from this by explicitly asking for the capabilities
5246 * using the Get_Source_Cap control message before falling back
5247 * to a hard reset. The control message should also be supported
5248 * and handled by all USB PD source and dual role devices
5249 * according to the specification.
5250 */
5251 if (tcpm_pd_send_control(port, PD_CTRL_GET_SOURCE_CAP, TCPC_TX_SOP))
5252 tcpm_set_state_cond(port, hard_reset_state(port), 0);
5253 else
5254 tcpm_set_state(port, hard_reset_state(port),
5255 port->timings.sink_wait_cap_time);
5256 break;
5257 case SNK_NEGOTIATE_CAPABILITIES:
5258 port->pd_capable = true;
5259 tcpm_set_partner_usb_comm_capable(port,
5260 !!(port->source_caps[0] & PDO_FIXED_USB_COMM));
5261 port->hard_reset_count = 0;
5262 ret = tcpm_pd_send_request(port);
5263 if (ret < 0) {
5264 /* Restore back to the original state */
5265 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_PD,
5266 port->pps_data.active,
5267 port->supply_voltage);
5268 /* Let the Source send capabilities again. */
5269 tcpm_set_state(port, SNK_WAIT_CAPABILITIES, 0);
5270 } else {
5271 tcpm_set_state_cond(port, hard_reset_state(port),
5272 PD_T_SENDER_RESPONSE);
5273 }
5274 break;
5275 case SNK_NEGOTIATE_PPS_CAPABILITIES:
5276 ret = tcpm_pd_send_pps_request(port);
5277 if (ret < 0) {
5278 /* Restore back to the original state */
5279 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_PD,
5280 port->pps_data.active,
5281 port->supply_voltage);
5282 port->pps_status = ret;
5283 /*
5284 * If this was called due to updates to sink
5285 * capabilities, and pps is no longer valid, we should
5286 * safely fall back to a standard PDO.
5287 */
5288 if (port->update_sink_caps)
5289 tcpm_set_state(port, SNK_NEGOTIATE_CAPABILITIES, 0);
5290 else
5291 tcpm_set_state(port, SNK_READY, 0);
5292 } else {
5293 tcpm_set_state_cond(port, hard_reset_state(port),
5294 PD_T_SENDER_RESPONSE);
5295 }
5296 break;
5297 case SNK_TRANSITION_SINK:
5298 /* From the USB PD spec:
5299 * "The Sink Shall transition to Sink Standby before a positive or
5300 * negative voltage transition of VBUS. During Sink Standby
5301 * the Sink Shall reduce its power draw to pSnkStdby."
5302 *
5303 * This is not applicable to PPS though as the port can continue
5304 * to draw negotiated power without switching to standby.
5305 */
5306 if (port->supply_voltage != port->req_supply_voltage && !port->pps_data.active &&
5307 port->current_limit * port->supply_voltage / 1000 > PD_P_SNK_STDBY_MW) {
5308 u32 stdby_ma = PD_P_SNK_STDBY_MW * 1000 / port->supply_voltage;
5309
5310 tcpm_log(port, "Setting standby current %u mV @ %u mA",
5311 port->supply_voltage, stdby_ma);
5312 tcpm_set_current_limit(port, stdby_ma, port->supply_voltage);
5313 }
5314 fallthrough;
5315 case SNK_TRANSITION_SINK_VBUS:
5316 tcpm_set_state(port, hard_reset_state(port),
5317 PD_T_PS_TRANSITION);
5318 break;
5319 case SNK_READY:
5320 port->try_snk_count = 0;
5321 port->update_sink_caps = false;
5322 if (port->explicit_contract) {
5323 typec_set_pwr_opmode(port->typec_port,
5324 TYPEC_PWR_MODE_PD);
5325 port->pwr_opmode = TYPEC_PWR_MODE_PD;
5326 }
5327
5328 if (!port->pd_capable && port->slow_charger_loop)
5329 tcpm_set_current_limit(port, tcpm_get_current_limit(port), 5000);
5330 tcpm_swap_complete(port, 0);
5331 tcpm_typec_connect(port);
5332 if (port->pd_capable && port->source_caps[0] & PDO_FIXED_DUAL_ROLE)
5333 mod_enable_frs_delayed_work(port, 0);
5334 tcpm_pps_complete(port, port->pps_status);
5335
5336 if (port->ams != NONE_AMS)
5337 tcpm_ams_finish(port);
5338 if (port->next_ams != NONE_AMS) {
5339 port->ams = port->next_ams;
5340 port->next_ams = NONE_AMS;
5341 }
5342
5343 /*
5344 * If previous AMS is interrupted, switch to the upcoming
5345 * state.
5346 */
5347 if (port->upcoming_state != INVALID_STATE) {
5348 upcoming_state = port->upcoming_state;
5349 port->upcoming_state = INVALID_STATE;
5350 tcpm_set_state(port, upcoming_state, 0);
5351 break;
5352 }
5353
5354 /*
5355 * 6.4.4.3.1 Discover Identity
5356 * "The Discover Identity Command Shall only be sent to SOP when there is an
5357 * Explicit Contract."
5358 *
5359 * Discover Identity on SOP' should be discovered prior to the
5360 * ready state, but if done after a Vconn Swap following Discover
5361 * Identity on SOP then the discovery process can be run here
5362 * as well.
5363 */
5364 if (port->explicit_contract) {
5365 if (port->send_discover_prime) {
5366 port->tx_sop_type = TCPC_TX_SOP_PRIME;
5367 } else {
5368 port->tx_sop_type = TCPC_TX_SOP;
5369 tcpm_set_initial_svdm_version(port);
5370 }
5371 mod_send_discover_delayed_work(port, 0);
5372 } else {
5373 port->send_discover = false;
5374 port->send_discover_prime = false;
5375 }
5376
5377 power_supply_changed(port->psy);
5378 break;
5379
5380 /* Accessory states */
5381 case ACC_UNATTACHED:
5382 tcpm_acc_detach(port);
5383 if (port->port_type == TYPEC_PORT_SRC)
5384 tcpm_set_state(port, SRC_UNATTACHED, 0);
5385 else
5386 tcpm_set_state(port, SNK_UNATTACHED, 0);
5387 break;
5388 case DEBUG_ACC_ATTACHED:
5389 case AUDIO_ACC_ATTACHED:
5390 ret = tcpm_acc_attach(port);
5391 if (ret < 0)
5392 tcpm_set_state(port, ACC_UNATTACHED, 0);
5393 break;
5394 case DEBUG_ACC_DEBOUNCE:
5395 case AUDIO_ACC_DEBOUNCE:
5396 tcpm_set_state(port, ACC_UNATTACHED, port->timings.cc_debounce_time);
5397 break;
5398
5399 /* Hard_Reset states */
5400 case HARD_RESET_SEND:
5401 if (port->ams != NONE_AMS)
5402 tcpm_ams_finish(port);
5403 if (!port->self_powered && port->port_type == TYPEC_PORT_SNK)
5404 dev_err(port->dev, "Initiating hard-reset, which might result in machine power-loss.\n");
5405 /*
5406 * State machine will be directed to HARD_RESET_START,
5407 * thus set upcoming_state to INVALID_STATE.
5408 */
5409 port->upcoming_state = INVALID_STATE;
5410 tcpm_ams_start(port, HARD_RESET);
5411 break;
5412 case HARD_RESET_START:
5413 port->sink_cap_done = false;
5414 if (port->tcpc->enable_frs)
5415 port->tcpc->enable_frs(port->tcpc, false);
5416 port->hard_reset_count++;
5417 port->tcpc->set_pd_rx(port->tcpc, false);
5418 tcpm_unregister_altmodes(port);
5419 port->nr_sink_caps = 0;
5420 port->send_discover = true;
5421 port->send_discover_prime = false;
5422 if (port->pwr_role == TYPEC_SOURCE)
5423 tcpm_set_state(port, SRC_HARD_RESET_VBUS_OFF,
5424 PD_T_PS_HARD_RESET);
5425 else
5426 tcpm_set_state(port, SNK_HARD_RESET_SINK_OFF, 0);
5427 break;
5428 case SRC_HARD_RESET_VBUS_OFF:
5429 /*
5430 * 7.1.5 Response to Hard Resets
5431 * Hard Reset Signaling indicates a communication failure has occurred and the
5432 * Source Shall stop driving VCONN, Shall remove Rp from the VCONN pin and Shall
5433 * drive VBUS to vSafe0V as shown in Figure 7-9.
5434 */
5435 tcpm_set_vconn(port, false);
5436 tcpm_set_vbus(port, false);
5437 tcpm_set_roles(port, port->self_powered, TYPEC_STATE_USB, TYPEC_SOURCE,
5438 tcpm_data_role_for_source(port));
5439 /*
5440 * If tcpc fails to notify vbus off, TCPM will wait for PD_T_SAFE_0V +
5441 * PD_T_SRC_RECOVER before turning vbus back on.
5442 * From Table 7-12 Sequence Description for a Source Initiated Hard Reset:
5443 * 4. Policy Engine waits tPSHardReset after sending Hard Reset Signaling and then
5444 * tells the Device Policy Manager to instruct the power supply to perform a
5445 * Hard Reset. The transition to vSafe0V Shall occur within tSafe0V (t2).
5446 * 5. After tSrcRecover the Source applies power to VBUS in an attempt to
5447 * re-establish communication with the Sink and resume USB Default Operation.
5448 * The transition to vSafe5V Shall occur within tSrcTurnOn(t4).
5449 */
5450 tcpm_set_state(port, SRC_HARD_RESET_VBUS_ON, PD_T_SAFE_0V + PD_T_SRC_RECOVER);
5451 break;
5452 case SRC_HARD_RESET_VBUS_ON:
5453 tcpm_set_vconn(port, true);
5454 tcpm_set_vbus(port, true);
5455 if (port->ams == HARD_RESET)
5456 tcpm_ams_finish(port);
5457 if (port->pd_supported)
5458 port->tcpc->set_pd_rx(port->tcpc, true);
5459 tcpm_set_attached_state(port, true);
5460 tcpm_set_state(port, SRC_UNATTACHED, PD_T_PS_SOURCE_ON);
5461 break;
5462 case SNK_HARD_RESET_SINK_OFF:
5463 /* Do not discharge/disconnect during hard reset */
5464 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_USB, false, 0);
5465 memset(&port->pps_data, 0, sizeof(port->pps_data));
5466 tcpm_set_vconn(port, false);
5467 if (port->pd_capable)
5468 tcpm_set_charge(port, false);
5469 tcpm_set_roles(port, port->self_powered, TYPEC_STATE_USB, TYPEC_SINK,
5470 tcpm_data_role_for_sink(port));
5471 /*
5472 * VBUS may or may not toggle, depending on the adapter.
5473 * If it doesn't toggle, transition to SNK_HARD_RESET_SINK_ON
5474 * directly after timeout.
5475 */
5476 tcpm_set_state(port, SNK_HARD_RESET_SINK_ON, PD_T_SAFE_0V);
5477 break;
5478 case SNK_HARD_RESET_WAIT_VBUS:
5479 if (port->ams == HARD_RESET)
5480 tcpm_ams_finish(port);
5481 /* Assume we're disconnected if VBUS doesn't come back. */
5482 tcpm_set_state(port, SNK_UNATTACHED,
5483 PD_T_SRC_RECOVER_MAX + PD_T_SRC_TURN_ON);
5484 break;
5485 case SNK_HARD_RESET_SINK_ON:
5486 /* Note: There is no guarantee that VBUS is on in this state */
5487 /*
5488 * XXX:
5489 * The specification suggests that dual mode ports in sink
5490 * mode should transition to state PE_SRC_Transition_to_default.
5491 * See USB power delivery specification chapter 8.3.3.6.1.3.
5492 * This would mean to
5493 * - turn off VCONN, reset power supply
5494 * - request hardware reset
5495 * - turn on VCONN
5496 * - Transition to state PE_Src_Startup
5497 * SNK only ports shall transition to state Snk_Startup
5498 * (see chapter 8.3.3.3.8).
5499 * Similar, dual-mode ports in source mode should transition
5500 * to PE_SNK_Transition_to_default.
5501 */
5502 if (port->pd_capable) {
5503 tcpm_set_current_limit(port,
5504 tcpm_get_current_limit(port),
5505 5000);
5506 /* Not sink vbus if operational current is 0mA */
5507 tcpm_set_charge(port, !!pdo_max_current(port->snk_pdo[0]));
5508 }
5509 if (port->ams == HARD_RESET)
5510 tcpm_ams_finish(port);
5511 tcpm_set_attached_state(port, true);
5512 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_USB, false, VSAFE5V);
5513 tcpm_set_state(port, SNK_STARTUP, 0);
5514 break;
5515
5516 /* Soft_Reset states */
5517 case SOFT_RESET:
5518 port->message_id = 0;
5519 port->rx_msgid = -1;
5520 /* remove existing capabilities */
5521 usb_power_delivery_unregister_capabilities(port->partner_source_caps);
5522 port->partner_source_caps = NULL;
5523 tcpm_pd_send_control(port, PD_CTRL_ACCEPT, TCPC_TX_SOP);
5524 tcpm_ams_finish(port);
5525 if (port->pwr_role == TYPEC_SOURCE) {
5526 port->upcoming_state = SRC_SEND_CAPABILITIES;
5527 tcpm_ams_start(port, POWER_NEGOTIATION);
5528 } else {
5529 tcpm_set_state(port, SNK_WAIT_CAPABILITIES, 0);
5530 }
5531 break;
5532 case SRC_SOFT_RESET_WAIT_SNK_TX:
5533 case SNK_SOFT_RESET:
5534 if (port->ams != NONE_AMS)
5535 tcpm_ams_finish(port);
5536 port->upcoming_state = SOFT_RESET_SEND;
5537 tcpm_ams_start(port, SOFT_RESET_AMS);
5538 break;
5539 case SOFT_RESET_SEND:
5540 /*
5541 * Power Delivery 3.0 Section 6.3.13
5542 *
5543 * A Soft_Reset Message Shall be targeted at a specific entity
5544 * depending on the type of SOP* packet used.
5545 */
5546 if (port->tx_sop_type == TCPC_TX_SOP_PRIME) {
5547 port->message_id_prime = 0;
5548 port->rx_msgid_prime = -1;
5549 tcpm_pd_send_control(port, PD_CTRL_SOFT_RESET, TCPC_TX_SOP_PRIME);
5550 tcpm_set_state_cond(port, ready_state(port), PD_T_SENDER_RESPONSE);
5551 } else {
5552 port->message_id = 0;
5553 port->rx_msgid = -1;
5554 /* remove existing capabilities */
5555 usb_power_delivery_unregister_capabilities(port->partner_source_caps);
5556 port->partner_source_caps = NULL;
5557 if (tcpm_pd_send_control(port, PD_CTRL_SOFT_RESET, TCPC_TX_SOP))
5558 tcpm_set_state_cond(port, hard_reset_state(port), 0);
5559 else
5560 tcpm_set_state_cond(port, hard_reset_state(port),
5561 PD_T_SENDER_RESPONSE);
5562 }
5563 break;
5564
5565 /* DR_Swap states */
5566 case DR_SWAP_SEND:
5567 tcpm_pd_send_control(port, PD_CTRL_DR_SWAP, TCPC_TX_SOP);
5568 if (port->data_role == TYPEC_DEVICE || port->negotiated_rev > PD_REV20) {
5569 port->send_discover = true;
5570 port->send_discover_prime = false;
5571 }
5572 tcpm_set_state_cond(port, DR_SWAP_SEND_TIMEOUT,
5573 PD_T_SENDER_RESPONSE);
5574 break;
5575 case DR_SWAP_ACCEPT:
5576 tcpm_pd_send_control(port, PD_CTRL_ACCEPT, TCPC_TX_SOP);
5577 if (port->data_role == TYPEC_DEVICE || port->negotiated_rev > PD_REV20) {
5578 port->send_discover = true;
5579 port->send_discover_prime = false;
5580 }
5581 tcpm_set_state_cond(port, DR_SWAP_CHANGE_DR, 0);
5582 break;
5583 case DR_SWAP_SEND_TIMEOUT:
5584 tcpm_swap_complete(port, -ETIMEDOUT);
5585 port->send_discover = false;
5586 port->send_discover_prime = false;
5587 tcpm_ams_finish(port);
5588 tcpm_set_state(port, ready_state(port), 0);
5589 break;
5590 case DR_SWAP_CHANGE_DR:
5591 tcpm_unregister_altmodes(port);
5592 if (port->data_role == TYPEC_HOST)
5593 tcpm_set_roles(port, true, TYPEC_STATE_USB, port->pwr_role,
5594 TYPEC_DEVICE);
5595 else
5596 tcpm_set_roles(port, true, TYPEC_STATE_USB, port->pwr_role,
5597 TYPEC_HOST);
5598 tcpm_ams_finish(port);
5599 tcpm_set_state(port, ready_state(port), 0);
5600 break;
5601
5602 case FR_SWAP_SEND:
5603 if (tcpm_pd_send_control(port, PD_CTRL_FR_SWAP, TCPC_TX_SOP)) {
5604 tcpm_set_state(port, ERROR_RECOVERY, 0);
5605 break;
5606 }
5607 tcpm_set_state_cond(port, FR_SWAP_SEND_TIMEOUT, PD_T_SENDER_RESPONSE);
5608 break;
5609 case FR_SWAP_SEND_TIMEOUT:
5610 tcpm_set_state(port, ERROR_RECOVERY, 0);
5611 break;
5612 case FR_SWAP_SNK_SRC_TRANSITION_TO_OFF:
5613 tcpm_set_state(port, ERROR_RECOVERY, port->timings.ps_src_off_time);
5614 break;
5615 case FR_SWAP_SNK_SRC_NEW_SINK_READY:
5616 if (port->vbus_source)
5617 tcpm_set_state(port, FR_SWAP_SNK_SRC_SOURCE_VBUS_APPLIED, 0);
5618 else
5619 tcpm_set_state(port, ERROR_RECOVERY, PD_T_RECEIVER_RESPONSE);
5620 break;
5621 case FR_SWAP_SNK_SRC_SOURCE_VBUS_APPLIED:
5622 tcpm_set_pwr_role(port, TYPEC_SOURCE);
5623 if (tcpm_pd_send_control(port, PD_CTRL_PS_RDY, TCPC_TX_SOP)) {
5624 tcpm_set_state(port, ERROR_RECOVERY, 0);
5625 break;
5626 }
5627 tcpm_set_cc(port, tcpm_rp_cc(port));
5628 tcpm_set_state(port, SRC_STARTUP, PD_T_SWAP_SRC_START);
5629 break;
5630
5631 /* PR_Swap states */
5632 case PR_SWAP_ACCEPT:
5633 tcpm_pd_send_control(port, PD_CTRL_ACCEPT, TCPC_TX_SOP);
5634 tcpm_set_state(port, PR_SWAP_START, 0);
5635 break;
5636 case PR_SWAP_SEND:
5637 tcpm_pd_send_control(port, PD_CTRL_PR_SWAP, TCPC_TX_SOP);
5638 tcpm_set_state_cond(port, PR_SWAP_SEND_TIMEOUT,
5639 PD_T_SENDER_RESPONSE);
5640 break;
5641 case PR_SWAP_SEND_TIMEOUT:
5642 tcpm_swap_complete(port, -ETIMEDOUT);
5643 tcpm_set_state(port, ready_state(port), 0);
5644 break;
5645 case PR_SWAP_START:
5646 tcpm_apply_rc(port);
5647 if (port->pwr_role == TYPEC_SOURCE)
5648 tcpm_set_state(port, PR_SWAP_SRC_SNK_TRANSITION_OFF,
5649 PD_T_SRC_TRANSITION);
5650 else
5651 tcpm_set_state(port, PR_SWAP_SNK_SRC_SINK_OFF, 0);
5652 break;
5653 case PR_SWAP_SRC_SNK_TRANSITION_OFF:
5654 /*
5655 * Prevent vbus discharge circuit from turning on during PR_SWAP
5656 * as this is not a disconnect.
5657 */
5658 tcpm_set_vbus(port, false);
5659 port->explicit_contract = false;
5660 /* allow time for Vbus discharge, must be < tSrcSwapStdby */
5661 tcpm_set_state(port, PR_SWAP_SRC_SNK_SOURCE_OFF,
5662 PD_T_SRCSWAPSTDBY);
5663 break;
5664 case PR_SWAP_SRC_SNK_SOURCE_OFF:
5665 tcpm_set_cc(port, TYPEC_CC_RD);
5666 /* allow CC debounce */
5667 tcpm_set_state(port, PR_SWAP_SRC_SNK_SOURCE_OFF_CC_DEBOUNCED,
5668 port->timings.cc_debounce_time);
5669 break;
5670 case PR_SWAP_SRC_SNK_SOURCE_OFF_CC_DEBOUNCED:
5671 /*
5672 * USB-PD standard, 6.2.1.4, Port Power Role:
5673 * "During the Power Role Swap Sequence, for the initial Source
5674 * Port, the Port Power Role field shall be set to Sink in the
5675 * PS_RDY Message indicating that the initial Source’s power
5676 * supply is turned off"
5677 */
5678 tcpm_set_pwr_role(port, TYPEC_SINK);
5679 if (tcpm_pd_send_control(port, PD_CTRL_PS_RDY, TCPC_TX_SOP)) {
5680 tcpm_set_state(port, ERROR_RECOVERY, 0);
5681 break;
5682 }
5683 tcpm_set_state(port, ERROR_RECOVERY, PD_T_PS_SOURCE_ON_PRS);
5684 break;
5685 case PR_SWAP_SRC_SNK_SINK_ON:
5686 tcpm_enable_auto_vbus_discharge(port, true);
5687 /* Set the vbus disconnect threshold for implicit contract */
5688 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_USB, false, VSAFE5V);
5689 tcpm_set_state(port, SNK_STARTUP, 0);
5690 break;
5691 case PR_SWAP_SNK_SRC_SINK_OFF:
5692 /* will be source, remove existing capabilities */
5693 usb_power_delivery_unregister_capabilities(port->partner_source_caps);
5694 port->partner_source_caps = NULL;
5695 /*
5696 * Prevent vbus discharge circuit from turning on during PR_SWAP
5697 * as this is not a disconnect.
5698 */
5699 tcpm_set_auto_vbus_discharge_threshold(port, TYPEC_PWR_MODE_USB,
5700 port->pps_data.active, 0);
5701 tcpm_set_charge(port, false);
5702 tcpm_set_state(port, ERROR_RECOVERY, port->timings.ps_src_off_time);
5703 break;
5704 case PR_SWAP_SNK_SRC_SOURCE_ON:
5705 tcpm_enable_auto_vbus_discharge(port, true);
5706 tcpm_set_cc(port, tcpm_rp_cc(port));
5707 tcpm_set_vbus(port, true);
5708 /*
5709 * allow time VBUS ramp-up, must be < tNewSrc
5710 * Also, this window overlaps with CC debounce as well.
5711 * So, Wait for the max of two which is PD_T_NEWSRC
5712 */
5713 tcpm_set_state(port, PR_SWAP_SNK_SRC_SOURCE_ON_VBUS_RAMPED_UP,
5714 PD_T_NEWSRC);
5715 break;
5716 case PR_SWAP_SNK_SRC_SOURCE_ON_VBUS_RAMPED_UP:
5717 /*
5718 * USB PD standard, 6.2.1.4:
5719 * "Subsequent Messages initiated by the Policy Engine,
5720 * such as the PS_RDY Message sent to indicate that Vbus
5721 * is ready, will have the Port Power Role field set to
5722 * Source."
5723 */
5724 tcpm_set_pwr_role(port, TYPEC_SOURCE);
5725 tcpm_pd_send_control(port, PD_CTRL_PS_RDY, TCPC_TX_SOP);
5726 tcpm_set_state(port, SRC_STARTUP, PD_T_SWAP_SRC_START);
5727 break;
5728
5729 case VCONN_SWAP_ACCEPT:
5730 tcpm_pd_send_control(port, PD_CTRL_ACCEPT, TCPC_TX_SOP);
5731 tcpm_ams_finish(port);
5732 tcpm_set_state(port, VCONN_SWAP_START, 0);
5733 break;
5734 case VCONN_SWAP_SEND:
5735 tcpm_pd_send_control(port, PD_CTRL_VCONN_SWAP, TCPC_TX_SOP);
5736 tcpm_set_state(port, VCONN_SWAP_SEND_TIMEOUT,
5737 PD_T_SENDER_RESPONSE);
5738 break;
5739 case VCONN_SWAP_SEND_TIMEOUT:
5740 tcpm_swap_complete(port, -ETIMEDOUT);
5741 tcpm_set_state(port, ready_state(port), 0);
5742 break;
5743 case VCONN_SWAP_START:
5744 if (port->vconn_role == TYPEC_SOURCE)
5745 tcpm_set_state(port, VCONN_SWAP_WAIT_FOR_VCONN, 0);
5746 else
5747 tcpm_set_state(port, VCONN_SWAP_TURN_ON_VCONN, 0);
5748 break;
5749 case VCONN_SWAP_WAIT_FOR_VCONN:
5750 tcpm_set_state(port, hard_reset_state(port),
5751 PD_T_VCONN_SOURCE_ON);
5752 break;
5753 case VCONN_SWAP_TURN_ON_VCONN:
5754 ret = tcpm_set_vconn(port, true);
5755 tcpm_pd_send_control(port, PD_CTRL_PS_RDY, TCPC_TX_SOP);
5756 /*
5757 * USB PD 3.0 Section 6.4.4.3.1
5758 *
5759 * Note that a Cable Plug or VPD will not be ready for PD
5760 * Communication until tVCONNStable after VCONN has been applied
5761 */
5762 if (!ret)
5763 tcpm_set_state(port, VCONN_SWAP_SEND_SOFT_RESET,
5764 PD_T_VCONN_STABLE);
5765 else
5766 tcpm_set_state(port, ready_state(port), 0);
5767 break;
5768 case VCONN_SWAP_TURN_OFF_VCONN:
5769 tcpm_set_vconn(port, false);
5770 tcpm_set_state(port, ready_state(port), 0);
5771 break;
5772 case VCONN_SWAP_SEND_SOFT_RESET:
5773 tcpm_swap_complete(port, port->swap_status);
5774 if (tcpm_can_communicate_sop_prime(port)) {
5775 port->tx_sop_type = TCPC_TX_SOP_PRIME;
5776 port->upcoming_state = SOFT_RESET_SEND;
5777 tcpm_ams_start(port, SOFT_RESET_AMS);
5778 } else {
5779 tcpm_set_state(port, ready_state(port), 0);
5780 }
5781 break;
5782
5783 case DR_SWAP_CANCEL:
5784 case PR_SWAP_CANCEL:
5785 case VCONN_SWAP_CANCEL:
5786 tcpm_swap_complete(port, port->swap_status);
5787 if (port->pwr_role == TYPEC_SOURCE)
5788 tcpm_set_state(port, SRC_READY, 0);
5789 else
5790 tcpm_set_state(port, SNK_READY, 0);
5791 break;
5792 case FR_SWAP_CANCEL:
5793 if (port->pwr_role == TYPEC_SOURCE)
5794 tcpm_set_state(port, SRC_READY, 0);
5795 else
5796 tcpm_set_state(port, SNK_READY, 0);
5797 break;
5798
5799 case BIST_RX:
5800 switch (BDO_MODE_MASK(port->bist_request)) {
5801 case BDO_MODE_CARRIER2:
5802 tcpm_pd_transmit(port, TCPC_TX_BIST_MODE_2, NULL);
5803 tcpm_set_state(port, unattached_state(port),
5804 PD_T_BIST_CONT_MODE);
5805 break;
5806 case BDO_MODE_TESTDATA:
5807 if (port->tcpc->set_bist_data) {
5808 tcpm_log(port, "Enable BIST MODE TESTDATA");
5809 port->tcpc->set_bist_data(port->tcpc, true);
5810 }
5811 break;
5812 default:
5813 break;
5814 }
5815 break;
5816 case GET_STATUS_SEND:
5817 tcpm_pd_send_control(port, PD_CTRL_GET_STATUS, TCPC_TX_SOP);
5818 tcpm_set_state(port, GET_STATUS_SEND_TIMEOUT,
5819 PD_T_SENDER_RESPONSE);
5820 break;
5821 case GET_STATUS_SEND_TIMEOUT:
5822 tcpm_set_state(port, ready_state(port), 0);
5823 break;
5824 case GET_PPS_STATUS_SEND:
5825 tcpm_pd_send_control(port, PD_CTRL_GET_PPS_STATUS, TCPC_TX_SOP);
5826 tcpm_set_state(port, GET_PPS_STATUS_SEND_TIMEOUT,
5827 PD_T_SENDER_RESPONSE);
5828 break;
5829 case GET_PPS_STATUS_SEND_TIMEOUT:
5830 tcpm_set_state(port, ready_state(port), 0);
5831 break;
5832 case GET_SINK_CAP:
5833 tcpm_pd_send_control(port, PD_CTRL_GET_SINK_CAP, TCPC_TX_SOP);
5834 tcpm_set_state(port, GET_SINK_CAP_TIMEOUT, PD_T_SENDER_RESPONSE);
5835 break;
5836 case GET_SINK_CAP_TIMEOUT:
5837 port->sink_cap_done = true;
5838 tcpm_set_state(port, ready_state(port), 0);
5839 break;
5840 case ERROR_RECOVERY:
5841 tcpm_swap_complete(port, -EPROTO);
5842 tcpm_pps_complete(port, -EPROTO);
5843 tcpm_set_state(port, PORT_RESET, 0);
5844 break;
5845 case PORT_RESET:
5846 tcpm_reset_port(port);
5847 if (port->self_powered)
5848 tcpm_set_cc(port, TYPEC_CC_OPEN);
5849 else
5850 tcpm_set_cc(port, tcpm_default_state(port) == SNK_UNATTACHED ?
5851 TYPEC_CC_RD : tcpm_rp_cc(port));
5852 tcpm_set_state(port, PORT_RESET_WAIT_OFF,
5853 PD_T_ERROR_RECOVERY);
5854 break;
5855 case PORT_RESET_WAIT_OFF:
5856 tcpm_set_state(port,
5857 tcpm_default_state(port),
5858 port->vbus_present ? port->timings.ps_src_off_time : 0);
5859 break;
5860
5861 /* AMS intermediate state */
5862 case AMS_START:
5863 if (port->upcoming_state == INVALID_STATE) {
5864 tcpm_set_state(port, port->pwr_role == TYPEC_SOURCE ?
5865 SRC_READY : SNK_READY, 0);
5866 break;
5867 }
5868
5869 upcoming_state = port->upcoming_state;
5870 port->upcoming_state = INVALID_STATE;
5871 tcpm_set_state(port, upcoming_state, 0);
5872 break;
5873
5874 /* Chunk state */
5875 case CHUNK_NOT_SUPP:
5876 tcpm_pd_send_control(port, PD_CTRL_NOT_SUPP, TCPC_TX_SOP);
5877 tcpm_set_state(port, port->pwr_role == TYPEC_SOURCE ? SRC_READY : SNK_READY, 0);
5878 break;
5879
5880 /* Cable states */
5881 case SRC_VDM_IDENTITY_REQUEST:
5882 port->send_discover_prime = true;
5883 port->tx_sop_type = TCPC_TX_SOP_PRIME;
5884 mod_send_discover_delayed_work(port, 0);
5885 port->upcoming_state = SRC_SEND_CAPABILITIES;
5886 break;
5887
5888 default:
5889 WARN(1, "Unexpected port state %d\n", port->state);
5890 break;
5891 }
5892 }
5893
tcpm_state_machine_work(struct kthread_work * work)5894 static void tcpm_state_machine_work(struct kthread_work *work)
5895 {
5896 struct tcpm_port *port = container_of(work, struct tcpm_port, state_machine);
5897 enum tcpm_state prev_state;
5898
5899 mutex_lock(&port->lock);
5900 port->state_machine_running = true;
5901
5902 if (port->queued_message && tcpm_send_queued_message(port))
5903 goto done;
5904
5905 /* If we were queued due to a delayed state change, update it now */
5906 if (port->delayed_state) {
5907 tcpm_log(port, "state change %s -> %s [delayed %ld ms]",
5908 tcpm_states[port->state],
5909 tcpm_states[port->delayed_state], port->delay_ms);
5910 port->prev_state = port->state;
5911 port->state = port->delayed_state;
5912 port->delayed_state = INVALID_STATE;
5913 }
5914
5915 /*
5916 * Continue running as long as we have (non-delayed) state changes
5917 * to make.
5918 */
5919 do {
5920 prev_state = port->state;
5921 run_state_machine(port);
5922 if (port->queued_message)
5923 tcpm_send_queued_message(port);
5924 } while (port->state != prev_state && !port->delayed_state);
5925
5926 done:
5927 port->state_machine_running = false;
5928 mutex_unlock(&port->lock);
5929 }
5930
_tcpm_cc_change(struct tcpm_port * port,enum typec_cc_status cc1,enum typec_cc_status cc2)5931 static void _tcpm_cc_change(struct tcpm_port *port, enum typec_cc_status cc1,
5932 enum typec_cc_status cc2)
5933 {
5934 enum typec_cc_status old_cc1, old_cc2;
5935 enum tcpm_state new_state;
5936
5937 old_cc1 = port->cc1;
5938 old_cc2 = port->cc2;
5939 port->cc1 = cc1;
5940 port->cc2 = cc2;
5941
5942 tcpm_log_force(port,
5943 "CC1: %u -> %u, CC2: %u -> %u [state %s, polarity %d, %s]",
5944 old_cc1, cc1, old_cc2, cc2, tcpm_states[port->state],
5945 port->polarity,
5946 tcpm_port_is_disconnected(port) ? "disconnected"
5947 : "connected");
5948
5949 switch (port->state) {
5950 case TOGGLING:
5951 if (tcpm_port_is_debug(port) || tcpm_port_is_audio(port) ||
5952 tcpm_port_is_source(port))
5953 tcpm_set_state(port, SRC_ATTACH_WAIT, 0);
5954 else if (tcpm_port_is_sink(port))
5955 tcpm_set_state(port, SNK_ATTACH_WAIT, 0);
5956 break;
5957 case CHECK_CONTAMINANT:
5958 /* Wait for Toggling to be resumed */
5959 break;
5960 case SRC_UNATTACHED:
5961 case ACC_UNATTACHED:
5962 if (tcpm_port_is_debug(port) || tcpm_port_is_audio(port) ||
5963 tcpm_port_is_source(port))
5964 tcpm_set_state(port, SRC_ATTACH_WAIT, 0);
5965 break;
5966 case SRC_ATTACH_WAIT:
5967 if (tcpm_port_is_disconnected(port) ||
5968 tcpm_port_is_audio_detached(port))
5969 tcpm_set_state(port, SRC_UNATTACHED, 0);
5970 else if (cc1 != old_cc1 || cc2 != old_cc2)
5971 tcpm_set_state(port, SRC_ATTACH_WAIT, 0);
5972 break;
5973 case SRC_ATTACHED:
5974 case SRC_STARTUP:
5975 case SRC_SEND_CAPABILITIES:
5976 case SRC_READY:
5977 if (tcpm_port_is_disconnected(port) ||
5978 !tcpm_port_is_source(port)) {
5979 if (port->port_type == TYPEC_PORT_SRC)
5980 tcpm_set_state(port, SRC_UNATTACHED, tcpm_wait_for_discharge(port));
5981 else
5982 tcpm_set_state(port, SNK_UNATTACHED, tcpm_wait_for_discharge(port));
5983 }
5984 break;
5985 case SNK_UNATTACHED:
5986 if (tcpm_port_is_debug(port) || tcpm_port_is_audio(port) ||
5987 tcpm_port_is_sink(port))
5988 tcpm_set_state(port, SNK_ATTACH_WAIT, 0);
5989 break;
5990 case SNK_ATTACH_WAIT:
5991 if ((port->cc1 == TYPEC_CC_OPEN &&
5992 port->cc2 != TYPEC_CC_OPEN) ||
5993 (port->cc1 != TYPEC_CC_OPEN &&
5994 port->cc2 == TYPEC_CC_OPEN))
5995 new_state = SNK_DEBOUNCED;
5996 else if (tcpm_port_is_disconnected(port))
5997 new_state = SNK_UNATTACHED;
5998 else
5999 break;
6000 if (new_state != port->delayed_state)
6001 tcpm_set_state(port, SNK_ATTACH_WAIT, 0);
6002 break;
6003 case SNK_DEBOUNCED:
6004 if (tcpm_port_is_disconnected(port))
6005 new_state = SNK_UNATTACHED;
6006 else if (port->vbus_present)
6007 new_state = tcpm_try_src(port) ? SRC_TRY : SNK_ATTACHED;
6008 else
6009 new_state = SNK_UNATTACHED;
6010 if (new_state != port->delayed_state)
6011 tcpm_set_state(port, SNK_DEBOUNCED, 0);
6012 break;
6013 case SNK_READY:
6014 /*
6015 * EXIT condition is based primarily on vbus disconnect and CC is secondary.
6016 * "A port that has entered into USB PD communications with the Source and
6017 * has seen the CC voltage exceed vRd-USB may monitor the CC pin to detect
6018 * cable disconnect in addition to monitoring VBUS.
6019 *
6020 * A port that is monitoring the CC voltage for disconnect (but is not in
6021 * the process of a USB PD PR_Swap or USB PD FR_Swap) shall transition to
6022 * Unattached.SNK within tSinkDisconnect after the CC voltage remains below
6023 * vRd-USB for tPDDebounce."
6024 *
6025 * When set_auto_vbus_discharge_threshold is enabled, CC pins go
6026 * away before vbus decays to disconnect threshold. Allow
6027 * disconnect to be driven by vbus disconnect when auto vbus
6028 * discharge is enabled.
6029 */
6030 if (!port->auto_vbus_discharge_enabled && tcpm_port_is_disconnected(port))
6031 tcpm_set_state(port, unattached_state(port), 0);
6032 else if (!port->pd_capable &&
6033 (cc1 != old_cc1 || cc2 != old_cc2))
6034 tcpm_set_current_limit(port,
6035 tcpm_get_current_limit(port),
6036 5000);
6037 break;
6038
6039 case AUDIO_ACC_ATTACHED:
6040 if (cc1 == TYPEC_CC_OPEN || cc2 == TYPEC_CC_OPEN)
6041 tcpm_set_state(port, AUDIO_ACC_DEBOUNCE, 0);
6042 break;
6043 case AUDIO_ACC_DEBOUNCE:
6044 if (tcpm_port_is_audio(port))
6045 tcpm_set_state(port, AUDIO_ACC_ATTACHED, 0);
6046 break;
6047
6048 case DEBUG_ACC_ATTACHED:
6049 if (cc1 == TYPEC_CC_OPEN || cc2 == TYPEC_CC_OPEN)
6050 tcpm_set_state(port, DEBUG_ACC_DEBOUNCE, 0);
6051 break;
6052
6053 case DEBUG_ACC_DEBOUNCE:
6054 if (tcpm_port_is_debug(port))
6055 tcpm_set_state(port, DEBUG_ACC_ATTACHED, 0);
6056 break;
6057
6058 case SNK_TRY:
6059 /* Do nothing, waiting for timeout */
6060 break;
6061
6062 case SNK_DISCOVERY:
6063 /* CC line is unstable, wait for debounce */
6064 if (tcpm_port_is_disconnected(port))
6065 tcpm_set_state(port, SNK_DISCOVERY_DEBOUNCE, 0);
6066 break;
6067 case SNK_DISCOVERY_DEBOUNCE:
6068 break;
6069
6070 case SRC_TRYWAIT:
6071 /* Hand over to state machine if needed */
6072 if (!port->vbus_present && tcpm_port_is_source(port))
6073 tcpm_set_state(port, SRC_TRYWAIT_DEBOUNCE, 0);
6074 break;
6075 case SRC_TRYWAIT_DEBOUNCE:
6076 if (port->vbus_present || !tcpm_port_is_source(port))
6077 tcpm_set_state(port, SRC_TRYWAIT, 0);
6078 break;
6079 case SNK_TRY_WAIT_DEBOUNCE:
6080 if (!tcpm_port_is_sink(port)) {
6081 port->max_wait = 0;
6082 tcpm_set_state(port, SRC_TRYWAIT, PD_T_PD_DEBOUNCE);
6083 }
6084 break;
6085 case SRC_TRY_WAIT:
6086 if (tcpm_port_is_source(port))
6087 tcpm_set_state(port, SRC_TRY_DEBOUNCE, 0);
6088 break;
6089 case SRC_TRY_DEBOUNCE:
6090 tcpm_set_state(port, SRC_TRY_WAIT, 0);
6091 break;
6092 case SNK_TRYWAIT_DEBOUNCE:
6093 if (tcpm_port_is_sink(port))
6094 tcpm_set_state(port, SNK_TRYWAIT_VBUS, 0);
6095 break;
6096 case SNK_TRYWAIT_VBUS:
6097 if (!tcpm_port_is_sink(port))
6098 tcpm_set_state(port, SNK_TRYWAIT_DEBOUNCE, 0);
6099 break;
6100 case SNK_TRY_WAIT_DEBOUNCE_CHECK_VBUS:
6101 if (!tcpm_port_is_sink(port))
6102 tcpm_set_state(port, SRC_TRYWAIT, PD_T_TRY_CC_DEBOUNCE);
6103 else
6104 tcpm_set_state(port, SNK_TRY_WAIT_DEBOUNCE_CHECK_VBUS, 0);
6105 break;
6106 case SNK_TRYWAIT:
6107 /* Do nothing, waiting for tCCDebounce */
6108 break;
6109 case PR_SWAP_SNK_SRC_SINK_OFF:
6110 case PR_SWAP_SRC_SNK_TRANSITION_OFF:
6111 case PR_SWAP_SRC_SNK_SOURCE_OFF:
6112 case PR_SWAP_SRC_SNK_SOURCE_OFF_CC_DEBOUNCED:
6113 case PR_SWAP_SNK_SRC_SOURCE_ON:
6114 /*
6115 * CC state change is expected in PR_SWAP
6116 * Ignore it.
6117 */
6118 break;
6119 case FR_SWAP_SEND:
6120 case FR_SWAP_SEND_TIMEOUT:
6121 case FR_SWAP_SNK_SRC_TRANSITION_TO_OFF:
6122 case FR_SWAP_SNK_SRC_NEW_SINK_READY:
6123 case FR_SWAP_SNK_SRC_SOURCE_VBUS_APPLIED:
6124 /* Do nothing, CC change expected */
6125 break;
6126
6127 case PORT_RESET:
6128 case PORT_RESET_WAIT_OFF:
6129 /*
6130 * State set back to default mode once the timer completes.
6131 * Ignore CC changes here.
6132 */
6133 break;
6134 default:
6135 /*
6136 * While acting as sink and auto vbus discharge is enabled, Allow disconnect
6137 * to be driven by vbus disconnect.
6138 */
6139 if (tcpm_port_is_disconnected(port) && !(port->pwr_role == TYPEC_SINK &&
6140 port->auto_vbus_discharge_enabled))
6141 tcpm_set_state(port, unattached_state(port), 0);
6142 break;
6143 }
6144 }
6145
_tcpm_pd_vbus_on(struct tcpm_port * port)6146 static void _tcpm_pd_vbus_on(struct tcpm_port *port)
6147 {
6148 tcpm_log_force(port, "VBUS on");
6149 port->vbus_present = true;
6150 /*
6151 * When vbus_present is true i.e. Voltage at VBUS is greater than VSAFE5V implicitly
6152 * states that vbus is not at VSAFE0V, hence clear the vbus_vsafe0v flag here.
6153 */
6154 port->vbus_vsafe0v = false;
6155
6156 switch (port->state) {
6157 case SNK_TRANSITION_SINK_VBUS:
6158 port->explicit_contract = true;
6159 tcpm_set_state(port, SNK_READY, 0);
6160 break;
6161 case SNK_DISCOVERY:
6162 tcpm_set_state(port, SNK_DISCOVERY, 0);
6163 break;
6164
6165 case SNK_DEBOUNCED:
6166 tcpm_set_state(port, tcpm_try_src(port) ? SRC_TRY
6167 : SNK_ATTACHED,
6168 0);
6169 break;
6170 case SNK_HARD_RESET_WAIT_VBUS:
6171 tcpm_set_state(port, SNK_HARD_RESET_SINK_ON, 0);
6172 break;
6173 case SRC_ATTACHED:
6174 tcpm_set_state(port, SRC_STARTUP, 0);
6175 break;
6176 case SRC_HARD_RESET_VBUS_ON:
6177 tcpm_set_state(port, SRC_STARTUP, 0);
6178 break;
6179
6180 case SNK_TRY:
6181 /* Do nothing, waiting for timeout */
6182 break;
6183 case SRC_TRYWAIT:
6184 /* Do nothing, Waiting for Rd to be detected */
6185 break;
6186 case SRC_TRYWAIT_DEBOUNCE:
6187 tcpm_set_state(port, SRC_TRYWAIT, 0);
6188 break;
6189 case SNK_TRY_WAIT_DEBOUNCE:
6190 /* Do nothing, waiting for PD_DEBOUNCE to do be done */
6191 break;
6192 case SNK_TRYWAIT:
6193 /* Do nothing, waiting for tCCDebounce */
6194 break;
6195 case SNK_TRYWAIT_VBUS:
6196 if (tcpm_port_is_sink(port))
6197 tcpm_set_state(port, SNK_ATTACHED, 0);
6198 break;
6199 case SNK_TRYWAIT_DEBOUNCE:
6200 /* Do nothing, waiting for Rp */
6201 break;
6202 case SNK_TRY_WAIT_DEBOUNCE_CHECK_VBUS:
6203 if (port->vbus_present && tcpm_port_is_sink(port))
6204 tcpm_set_state(port, SNK_ATTACHED, 0);
6205 break;
6206 case SRC_TRY_WAIT:
6207 case SRC_TRY_DEBOUNCE:
6208 /* Do nothing, waiting for sink detection */
6209 break;
6210 case FR_SWAP_SEND:
6211 case FR_SWAP_SEND_TIMEOUT:
6212 case FR_SWAP_SNK_SRC_TRANSITION_TO_OFF:
6213 case FR_SWAP_SNK_SRC_SOURCE_VBUS_APPLIED:
6214 if (port->tcpc->frs_sourcing_vbus)
6215 port->tcpc->frs_sourcing_vbus(port->tcpc);
6216 break;
6217 case FR_SWAP_SNK_SRC_NEW_SINK_READY:
6218 if (port->tcpc->frs_sourcing_vbus)
6219 port->tcpc->frs_sourcing_vbus(port->tcpc);
6220 tcpm_set_state(port, FR_SWAP_SNK_SRC_SOURCE_VBUS_APPLIED, 0);
6221 break;
6222
6223 case PORT_RESET:
6224 case PORT_RESET_WAIT_OFF:
6225 /*
6226 * State set back to default mode once the timer completes.
6227 * Ignore vbus changes here.
6228 */
6229 break;
6230
6231 default:
6232 break;
6233 }
6234 }
6235
_tcpm_pd_vbus_off(struct tcpm_port * port)6236 static void _tcpm_pd_vbus_off(struct tcpm_port *port)
6237 {
6238 tcpm_log_force(port, "VBUS off");
6239 port->vbus_present = false;
6240 port->vbus_never_low = false;
6241 switch (port->state) {
6242 case SNK_HARD_RESET_SINK_OFF:
6243 tcpm_set_state(port, SNK_HARD_RESET_WAIT_VBUS, 0);
6244 break;
6245 case HARD_RESET_SEND:
6246 break;
6247 case SNK_TRY:
6248 /* Do nothing, waiting for timeout */
6249 break;
6250 case SRC_TRYWAIT:
6251 /* Hand over to state machine if needed */
6252 if (tcpm_port_is_source(port))
6253 tcpm_set_state(port, SRC_TRYWAIT_DEBOUNCE, 0);
6254 break;
6255 case SNK_TRY_WAIT_DEBOUNCE:
6256 /* Do nothing, waiting for PD_DEBOUNCE to do be done */
6257 break;
6258 case SNK_TRYWAIT:
6259 case SNK_TRYWAIT_VBUS:
6260 case SNK_TRYWAIT_DEBOUNCE:
6261 break;
6262 case SNK_ATTACH_WAIT:
6263 case SNK_DEBOUNCED:
6264 /* Do nothing, as TCPM is still waiting for vbus to reach VSAFE5V to connect */
6265 break;
6266
6267 case SNK_NEGOTIATE_CAPABILITIES:
6268 break;
6269
6270 case PR_SWAP_SRC_SNK_TRANSITION_OFF:
6271 tcpm_set_state(port, PR_SWAP_SRC_SNK_SOURCE_OFF, 0);
6272 break;
6273
6274 case PR_SWAP_SNK_SRC_SINK_OFF:
6275 /* Do nothing, expected */
6276 break;
6277
6278 case PR_SWAP_SNK_SRC_SOURCE_ON:
6279 /*
6280 * Do nothing when vbus off notification is received.
6281 * TCPM can wait for PD_T_NEWSRC in PR_SWAP_SNK_SRC_SOURCE_ON
6282 * for the vbus source to ramp up.
6283 */
6284 break;
6285
6286 case PORT_RESET_WAIT_OFF:
6287 tcpm_set_state(port, tcpm_default_state(port), 0);
6288 break;
6289
6290 case SRC_TRY_WAIT:
6291 case SRC_TRY_DEBOUNCE:
6292 /* Do nothing, waiting for sink detection */
6293 break;
6294
6295 case SRC_STARTUP:
6296 case SRC_SEND_CAPABILITIES:
6297 case SRC_SEND_CAPABILITIES_TIMEOUT:
6298 case SRC_NEGOTIATE_CAPABILITIES:
6299 case SRC_TRANSITION_SUPPLY:
6300 case SRC_READY:
6301 case SRC_WAIT_NEW_CAPABILITIES:
6302 /*
6303 * Force to unattached state to re-initiate connection.
6304 * DRP port should move to Unattached.SNK instead of Unattached.SRC if
6305 * sink removed. Although sink removal here is due to source's vbus collapse,
6306 * treat it the same way for consistency.
6307 */
6308 if (port->port_type == TYPEC_PORT_SRC)
6309 tcpm_set_state(port, SRC_UNATTACHED, tcpm_wait_for_discharge(port));
6310 else
6311 tcpm_set_state(port, SNK_UNATTACHED, tcpm_wait_for_discharge(port));
6312 break;
6313
6314 case PORT_RESET:
6315 /*
6316 * State set back to default mode once the timer completes.
6317 * Ignore vbus changes here.
6318 */
6319 break;
6320
6321 case FR_SWAP_SEND:
6322 case FR_SWAP_SEND_TIMEOUT:
6323 case FR_SWAP_SNK_SRC_TRANSITION_TO_OFF:
6324 case FR_SWAP_SNK_SRC_NEW_SINK_READY:
6325 case FR_SWAP_SNK_SRC_SOURCE_VBUS_APPLIED:
6326 /* Do nothing, vbus drop expected */
6327 break;
6328
6329 case SNK_HARD_RESET_WAIT_VBUS:
6330 /* Do nothing, its OK to receive vbus off events */
6331 break;
6332
6333 default:
6334 if (port->pwr_role == TYPEC_SINK && port->attached)
6335 tcpm_set_state(port, SNK_UNATTACHED, tcpm_wait_for_discharge(port));
6336 break;
6337 }
6338 }
6339
_tcpm_pd_vbus_vsafe0v(struct tcpm_port * port)6340 static void _tcpm_pd_vbus_vsafe0v(struct tcpm_port *port)
6341 {
6342 tcpm_log_force(port, "VBUS VSAFE0V");
6343 port->vbus_vsafe0v = true;
6344 switch (port->state) {
6345 case SRC_HARD_RESET_VBUS_OFF:
6346 /*
6347 * After establishing the vSafe0V voltage condition on VBUS, the Source Shall wait
6348 * tSrcRecover before re-applying VCONN and restoring VBUS to vSafe5V.
6349 */
6350 tcpm_set_state(port, SRC_HARD_RESET_VBUS_ON, PD_T_SRC_RECOVER);
6351 break;
6352 case SRC_ATTACH_WAIT:
6353 if (tcpm_port_is_source(port))
6354 tcpm_set_state(port, tcpm_try_snk(port) ? SNK_TRY : SRC_ATTACHED,
6355 port->timings.cc_debounce_time);
6356 break;
6357 case SRC_STARTUP:
6358 case SRC_SEND_CAPABILITIES:
6359 case SRC_SEND_CAPABILITIES_TIMEOUT:
6360 case SRC_NEGOTIATE_CAPABILITIES:
6361 case SRC_TRANSITION_SUPPLY:
6362 case SRC_READY:
6363 case SRC_WAIT_NEW_CAPABILITIES:
6364 if (port->auto_vbus_discharge_enabled) {
6365 if (port->port_type == TYPEC_PORT_SRC)
6366 tcpm_set_state(port, SRC_UNATTACHED, 0);
6367 else
6368 tcpm_set_state(port, SNK_UNATTACHED, 0);
6369 }
6370 break;
6371 case PR_SWAP_SNK_SRC_SINK_OFF:
6372 case PR_SWAP_SNK_SRC_SOURCE_ON:
6373 /* Do nothing, vsafe0v is expected during transition */
6374 break;
6375 case SNK_ATTACH_WAIT:
6376 case SNK_DEBOUNCED:
6377 /*Do nothing, still waiting for VSAFE5V for connect */
6378 break;
6379 case SNK_HARD_RESET_WAIT_VBUS:
6380 /* Do nothing, its OK to receive vbus off events */
6381 break;
6382 default:
6383 if (port->pwr_role == TYPEC_SINK && port->auto_vbus_discharge_enabled)
6384 tcpm_set_state(port, SNK_UNATTACHED, 0);
6385 break;
6386 }
6387 }
6388
_tcpm_pd_hard_reset(struct tcpm_port * port)6389 static void _tcpm_pd_hard_reset(struct tcpm_port *port)
6390 {
6391 tcpm_log_force(port, "Received hard reset");
6392 if (port->bist_request == BDO_MODE_TESTDATA && port->tcpc->set_bist_data)
6393 port->tcpc->set_bist_data(port->tcpc, false);
6394
6395 switch (port->state) {
6396 case TOGGLING:
6397 case ERROR_RECOVERY:
6398 case PORT_RESET:
6399 case PORT_RESET_WAIT_OFF:
6400 return;
6401 default:
6402 break;
6403 }
6404
6405 if (port->ams != NONE_AMS)
6406 port->ams = NONE_AMS;
6407 if (port->hard_reset_count < PD_N_HARD_RESET_COUNT)
6408 port->ams = HARD_RESET;
6409 /*
6410 * If we keep receiving hard reset requests, executing the hard reset
6411 * must have failed. Revert to error recovery if that happens.
6412 */
6413 tcpm_set_state(port,
6414 port->hard_reset_count < PD_N_HARD_RESET_COUNT ?
6415 HARD_RESET_START : ERROR_RECOVERY,
6416 0);
6417 }
6418
tcpm_pd_event_handler(struct kthread_work * work)6419 static void tcpm_pd_event_handler(struct kthread_work *work)
6420 {
6421 struct tcpm_port *port = container_of(work, struct tcpm_port,
6422 event_work);
6423 u32 events;
6424
6425 mutex_lock(&port->lock);
6426
6427 spin_lock(&port->pd_event_lock);
6428 while (port->pd_events) {
6429 events = port->pd_events;
6430 port->pd_events = 0;
6431 spin_unlock(&port->pd_event_lock);
6432 if (events & TCPM_RESET_EVENT)
6433 _tcpm_pd_hard_reset(port);
6434 if (events & TCPM_VBUS_EVENT) {
6435 bool vbus;
6436
6437 vbus = port->tcpc->get_vbus(port->tcpc);
6438 if (vbus) {
6439 _tcpm_pd_vbus_on(port);
6440 } else {
6441 _tcpm_pd_vbus_off(port);
6442 /*
6443 * When TCPC does not support detecting vsafe0v voltage level,
6444 * treat vbus absent as vsafe0v. Else invoke is_vbus_vsafe0v
6445 * to see if vbus has discharge to VSAFE0V.
6446 */
6447 if (!port->tcpc->is_vbus_vsafe0v ||
6448 port->tcpc->is_vbus_vsafe0v(port->tcpc))
6449 _tcpm_pd_vbus_vsafe0v(port);
6450 }
6451 }
6452 if (events & TCPM_CC_EVENT) {
6453 enum typec_cc_status cc1, cc2;
6454
6455 if (port->tcpc->get_cc(port->tcpc, &cc1, &cc2) == 0)
6456 _tcpm_cc_change(port, cc1, cc2);
6457 }
6458 if (events & TCPM_FRS_EVENT) {
6459 if (port->state == SNK_READY) {
6460 int ret;
6461
6462 port->upcoming_state = FR_SWAP_SEND;
6463 ret = tcpm_ams_start(port, FAST_ROLE_SWAP);
6464 if (ret == -EAGAIN)
6465 port->upcoming_state = INVALID_STATE;
6466 } else {
6467 tcpm_log(port, "Discarding FRS_SIGNAL! Not in sink ready");
6468 }
6469 }
6470 if (events & TCPM_SOURCING_VBUS) {
6471 tcpm_log(port, "sourcing vbus");
6472 /*
6473 * In fast role swap case TCPC autonomously sources vbus. Set vbus_source
6474 * true as TCPM wouldn't have called tcpm_set_vbus.
6475 *
6476 * When vbus is sourced on the command on TCPM i.e. TCPM called
6477 * tcpm_set_vbus to source vbus, vbus_source would already be true.
6478 */
6479 port->vbus_source = true;
6480 _tcpm_pd_vbus_on(port);
6481 }
6482 if (events & TCPM_PORT_CLEAN) {
6483 tcpm_log(port, "port clean");
6484 if (port->state == CHECK_CONTAMINANT) {
6485 if (tcpm_start_toggling(port, tcpm_rp_cc(port)))
6486 tcpm_set_state(port, TOGGLING, 0);
6487 else
6488 tcpm_set_state(port, tcpm_default_state(port), 0);
6489 }
6490 }
6491 if (events & TCPM_PORT_ERROR) {
6492 tcpm_log(port, "port triggering error recovery");
6493 tcpm_set_state(port, ERROR_RECOVERY, 0);
6494 }
6495
6496 spin_lock(&port->pd_event_lock);
6497 }
6498 spin_unlock(&port->pd_event_lock);
6499 mutex_unlock(&port->lock);
6500 }
6501
tcpm_cc_change(struct tcpm_port * port)6502 void tcpm_cc_change(struct tcpm_port *port)
6503 {
6504 spin_lock(&port->pd_event_lock);
6505 port->pd_events |= TCPM_CC_EVENT;
6506 spin_unlock(&port->pd_event_lock);
6507 kthread_queue_work(port->wq, &port->event_work);
6508 }
6509 EXPORT_SYMBOL_GPL(tcpm_cc_change);
6510
tcpm_vbus_change(struct tcpm_port * port)6511 void tcpm_vbus_change(struct tcpm_port *port)
6512 {
6513 spin_lock(&port->pd_event_lock);
6514 port->pd_events |= TCPM_VBUS_EVENT;
6515 spin_unlock(&port->pd_event_lock);
6516 kthread_queue_work(port->wq, &port->event_work);
6517 }
6518 EXPORT_SYMBOL_GPL(tcpm_vbus_change);
6519
tcpm_pd_hard_reset(struct tcpm_port * port)6520 void tcpm_pd_hard_reset(struct tcpm_port *port)
6521 {
6522 spin_lock(&port->pd_event_lock);
6523 port->pd_events = TCPM_RESET_EVENT;
6524 spin_unlock(&port->pd_event_lock);
6525 kthread_queue_work(port->wq, &port->event_work);
6526 }
6527 EXPORT_SYMBOL_GPL(tcpm_pd_hard_reset);
6528
tcpm_sink_frs(struct tcpm_port * port)6529 void tcpm_sink_frs(struct tcpm_port *port)
6530 {
6531 spin_lock(&port->pd_event_lock);
6532 port->pd_events |= TCPM_FRS_EVENT;
6533 spin_unlock(&port->pd_event_lock);
6534 kthread_queue_work(port->wq, &port->event_work);
6535 }
6536 EXPORT_SYMBOL_GPL(tcpm_sink_frs);
6537
tcpm_sourcing_vbus(struct tcpm_port * port)6538 void tcpm_sourcing_vbus(struct tcpm_port *port)
6539 {
6540 spin_lock(&port->pd_event_lock);
6541 port->pd_events |= TCPM_SOURCING_VBUS;
6542 spin_unlock(&port->pd_event_lock);
6543 kthread_queue_work(port->wq, &port->event_work);
6544 }
6545 EXPORT_SYMBOL_GPL(tcpm_sourcing_vbus);
6546
tcpm_port_clean(struct tcpm_port * port)6547 void tcpm_port_clean(struct tcpm_port *port)
6548 {
6549 spin_lock(&port->pd_event_lock);
6550 port->pd_events |= TCPM_PORT_CLEAN;
6551 spin_unlock(&port->pd_event_lock);
6552 kthread_queue_work(port->wq, &port->event_work);
6553 }
6554 EXPORT_SYMBOL_GPL(tcpm_port_clean);
6555
tcpm_port_is_toggling(struct tcpm_port * port)6556 bool tcpm_port_is_toggling(struct tcpm_port *port)
6557 {
6558 return port->port_type == TYPEC_PORT_DRP && port->state == TOGGLING;
6559 }
6560 EXPORT_SYMBOL_GPL(tcpm_port_is_toggling);
6561
tcpm_port_error_recovery(struct tcpm_port * port)6562 void tcpm_port_error_recovery(struct tcpm_port *port)
6563 {
6564 spin_lock(&port->pd_event_lock);
6565 port->pd_events |= TCPM_PORT_ERROR;
6566 spin_unlock(&port->pd_event_lock);
6567 kthread_queue_work(port->wq, &port->event_work);
6568 }
6569 EXPORT_SYMBOL_GPL(tcpm_port_error_recovery);
6570
tcpm_enable_frs_work(struct kthread_work * work)6571 static void tcpm_enable_frs_work(struct kthread_work *work)
6572 {
6573 struct tcpm_port *port = container_of(work, struct tcpm_port, enable_frs);
6574 int ret;
6575
6576 mutex_lock(&port->lock);
6577 /* Not FRS capable */
6578 if (!port->connected || port->port_type != TYPEC_PORT_DRP ||
6579 port->pwr_opmode != TYPEC_PWR_MODE_PD ||
6580 !port->tcpc->enable_frs ||
6581 /* Sink caps queried */
6582 port->sink_cap_done || port->negotiated_rev < PD_REV30)
6583 goto unlock;
6584
6585 /* Send when the state machine is idle */
6586 if (port->state != SNK_READY || port->vdm_sm_running || port->send_discover ||
6587 port->send_discover_prime)
6588 goto resched;
6589
6590 port->upcoming_state = GET_SINK_CAP;
6591 ret = tcpm_ams_start(port, GET_SINK_CAPABILITIES);
6592 if (ret == -EAGAIN) {
6593 port->upcoming_state = INVALID_STATE;
6594 } else {
6595 port->sink_cap_done = true;
6596 goto unlock;
6597 }
6598 resched:
6599 mod_enable_frs_delayed_work(port, GET_SINK_CAP_RETRY_MS);
6600 unlock:
6601 mutex_unlock(&port->lock);
6602 }
6603
tcpm_send_discover_work(struct kthread_work * work)6604 static void tcpm_send_discover_work(struct kthread_work *work)
6605 {
6606 struct tcpm_port *port = container_of(work, struct tcpm_port, send_discover_work);
6607
6608 mutex_lock(&port->lock);
6609 /* No need to send DISCOVER_IDENTITY anymore */
6610 if (!port->send_discover && !port->send_discover_prime)
6611 goto unlock;
6612
6613 if (port->data_role == TYPEC_DEVICE && port->negotiated_rev < PD_REV30) {
6614 port->send_discover = false;
6615 port->send_discover_prime = false;
6616 goto unlock;
6617 }
6618
6619 /* Retry if the port is not idle */
6620 if ((port->state != SRC_READY && port->state != SNK_READY &&
6621 port->state != SRC_VDM_IDENTITY_REQUEST) || port->vdm_sm_running) {
6622 mod_send_discover_delayed_work(port, SEND_DISCOVER_RETRY_MS);
6623 goto unlock;
6624 }
6625
6626 tcpm_send_vdm(port, USB_SID_PD, CMD_DISCOVER_IDENT, NULL, 0, port->tx_sop_type);
6627
6628 unlock:
6629 mutex_unlock(&port->lock);
6630 }
6631
tcpm_dr_set(struct typec_port * p,enum typec_data_role data)6632 static int tcpm_dr_set(struct typec_port *p, enum typec_data_role data)
6633 {
6634 struct tcpm_port *port = typec_get_drvdata(p);
6635 int ret;
6636
6637 mutex_lock(&port->swap_lock);
6638 mutex_lock(&port->lock);
6639
6640 if (port->typec_caps.data != TYPEC_PORT_DRD) {
6641 ret = -EINVAL;
6642 goto port_unlock;
6643 }
6644 if (port->state != SRC_READY && port->state != SNK_READY) {
6645 ret = -EAGAIN;
6646 goto port_unlock;
6647 }
6648
6649 if (port->data_role == data) {
6650 ret = 0;
6651 goto port_unlock;
6652 }
6653
6654 /*
6655 * XXX
6656 * 6.3.9: If an alternate mode is active, a request to swap
6657 * alternate modes shall trigger a port reset.
6658 * Reject data role swap request in this case.
6659 */
6660
6661 if (!port->pd_capable) {
6662 /*
6663 * If the partner is not PD capable, reset the port to
6664 * trigger a role change. This can only work if a preferred
6665 * role is configured, and if it matches the requested role.
6666 */
6667 if (port->try_role == TYPEC_NO_PREFERRED_ROLE ||
6668 port->try_role == port->pwr_role) {
6669 ret = -EINVAL;
6670 goto port_unlock;
6671 }
6672 port->non_pd_role_swap = true;
6673 tcpm_set_state(port, PORT_RESET, 0);
6674 } else {
6675 port->upcoming_state = DR_SWAP_SEND;
6676 ret = tcpm_ams_start(port, DATA_ROLE_SWAP);
6677 if (ret == -EAGAIN) {
6678 port->upcoming_state = INVALID_STATE;
6679 goto port_unlock;
6680 }
6681 }
6682
6683 port->swap_status = 0;
6684 port->swap_pending = true;
6685 reinit_completion(&port->swap_complete);
6686 mutex_unlock(&port->lock);
6687
6688 if (!wait_for_completion_timeout(&port->swap_complete,
6689 msecs_to_jiffies(PD_ROLE_SWAP_TIMEOUT)))
6690 ret = -ETIMEDOUT;
6691 else
6692 ret = port->swap_status;
6693
6694 port->non_pd_role_swap = false;
6695 goto swap_unlock;
6696
6697 port_unlock:
6698 mutex_unlock(&port->lock);
6699 swap_unlock:
6700 mutex_unlock(&port->swap_lock);
6701 return ret;
6702 }
6703
tcpm_pr_set(struct typec_port * p,enum typec_role role)6704 static int tcpm_pr_set(struct typec_port *p, enum typec_role role)
6705 {
6706 struct tcpm_port *port = typec_get_drvdata(p);
6707 int ret;
6708
6709 mutex_lock(&port->swap_lock);
6710 mutex_lock(&port->lock);
6711
6712 if (port->port_type != TYPEC_PORT_DRP) {
6713 ret = -EINVAL;
6714 goto port_unlock;
6715 }
6716 if (port->state != SRC_READY && port->state != SNK_READY) {
6717 ret = -EAGAIN;
6718 goto port_unlock;
6719 }
6720
6721 if (role == port->pwr_role) {
6722 ret = 0;
6723 goto port_unlock;
6724 }
6725
6726 port->upcoming_state = PR_SWAP_SEND;
6727 ret = tcpm_ams_start(port, POWER_ROLE_SWAP);
6728 if (ret == -EAGAIN) {
6729 port->upcoming_state = INVALID_STATE;
6730 goto port_unlock;
6731 }
6732
6733 port->swap_status = 0;
6734 port->swap_pending = true;
6735 reinit_completion(&port->swap_complete);
6736 mutex_unlock(&port->lock);
6737
6738 if (!wait_for_completion_timeout(&port->swap_complete,
6739 msecs_to_jiffies(PD_ROLE_SWAP_TIMEOUT)))
6740 ret = -ETIMEDOUT;
6741 else
6742 ret = port->swap_status;
6743
6744 goto swap_unlock;
6745
6746 port_unlock:
6747 mutex_unlock(&port->lock);
6748 swap_unlock:
6749 mutex_unlock(&port->swap_lock);
6750 return ret;
6751 }
6752
tcpm_vconn_set(struct typec_port * p,enum typec_role role)6753 static int tcpm_vconn_set(struct typec_port *p, enum typec_role role)
6754 {
6755 struct tcpm_port *port = typec_get_drvdata(p);
6756 int ret;
6757
6758 mutex_lock(&port->swap_lock);
6759 mutex_lock(&port->lock);
6760
6761 if (port->state != SRC_READY && port->state != SNK_READY) {
6762 ret = -EAGAIN;
6763 goto port_unlock;
6764 }
6765
6766 if (role == port->vconn_role) {
6767 ret = 0;
6768 goto port_unlock;
6769 }
6770
6771 port->upcoming_state = VCONN_SWAP_SEND;
6772 ret = tcpm_ams_start(port, VCONN_SWAP);
6773 if (ret == -EAGAIN) {
6774 port->upcoming_state = INVALID_STATE;
6775 goto port_unlock;
6776 }
6777
6778 port->swap_status = 0;
6779 port->swap_pending = true;
6780 reinit_completion(&port->swap_complete);
6781 mutex_unlock(&port->lock);
6782
6783 if (!wait_for_completion_timeout(&port->swap_complete,
6784 msecs_to_jiffies(PD_ROLE_SWAP_TIMEOUT)))
6785 ret = -ETIMEDOUT;
6786 else
6787 ret = port->swap_status;
6788
6789 goto swap_unlock;
6790
6791 port_unlock:
6792 mutex_unlock(&port->lock);
6793 swap_unlock:
6794 mutex_unlock(&port->swap_lock);
6795 return ret;
6796 }
6797
tcpm_try_role(struct typec_port * p,int role)6798 static int tcpm_try_role(struct typec_port *p, int role)
6799 {
6800 struct tcpm_port *port = typec_get_drvdata(p);
6801 struct tcpc_dev *tcpc = port->tcpc;
6802 int ret = 0;
6803
6804 mutex_lock(&port->lock);
6805 if (tcpc->try_role)
6806 ret = tcpc->try_role(tcpc, role);
6807 if (!ret)
6808 port->try_role = role;
6809 port->try_src_count = 0;
6810 port->try_snk_count = 0;
6811 mutex_unlock(&port->lock);
6812
6813 return ret;
6814 }
6815
tcpm_pps_set_op_curr(struct tcpm_port * port,u16 req_op_curr)6816 static int tcpm_pps_set_op_curr(struct tcpm_port *port, u16 req_op_curr)
6817 {
6818 unsigned int target_mw;
6819 int ret;
6820
6821 mutex_lock(&port->swap_lock);
6822 mutex_lock(&port->lock);
6823
6824 if (!port->pps_data.active) {
6825 ret = -EOPNOTSUPP;
6826 goto port_unlock;
6827 }
6828
6829 if (port->state != SNK_READY) {
6830 ret = -EAGAIN;
6831 goto port_unlock;
6832 }
6833
6834 if (req_op_curr > port->pps_data.max_curr) {
6835 ret = -EINVAL;
6836 goto port_unlock;
6837 }
6838
6839 target_mw = (req_op_curr * port->supply_voltage) / 1000;
6840 if (target_mw < port->operating_snk_mw) {
6841 ret = -EINVAL;
6842 goto port_unlock;
6843 }
6844
6845 port->upcoming_state = SNK_NEGOTIATE_PPS_CAPABILITIES;
6846 ret = tcpm_ams_start(port, POWER_NEGOTIATION);
6847 if (ret == -EAGAIN) {
6848 port->upcoming_state = INVALID_STATE;
6849 goto port_unlock;
6850 }
6851
6852 /* Round down operating current to align with PPS valid steps */
6853 req_op_curr = req_op_curr - (req_op_curr % RDO_PROG_CURR_MA_STEP);
6854
6855 reinit_completion(&port->pps_complete);
6856 port->pps_data.req_op_curr = req_op_curr;
6857 port->pps_status = 0;
6858 port->pps_pending = true;
6859 mutex_unlock(&port->lock);
6860
6861 if (!wait_for_completion_timeout(&port->pps_complete,
6862 msecs_to_jiffies(PD_PPS_CTRL_TIMEOUT)))
6863 ret = -ETIMEDOUT;
6864 else
6865 ret = port->pps_status;
6866
6867 goto swap_unlock;
6868
6869 port_unlock:
6870 mutex_unlock(&port->lock);
6871 swap_unlock:
6872 mutex_unlock(&port->swap_lock);
6873
6874 return ret;
6875 }
6876
tcpm_pps_set_out_volt(struct tcpm_port * port,u16 req_out_volt)6877 static int tcpm_pps_set_out_volt(struct tcpm_port *port, u16 req_out_volt)
6878 {
6879 unsigned int target_mw;
6880 int ret;
6881
6882 mutex_lock(&port->swap_lock);
6883 mutex_lock(&port->lock);
6884
6885 if (!port->pps_data.active) {
6886 ret = -EOPNOTSUPP;
6887 goto port_unlock;
6888 }
6889
6890 if (port->state != SNK_READY) {
6891 ret = -EAGAIN;
6892 goto port_unlock;
6893 }
6894
6895 target_mw = (port->current_limit * req_out_volt) / 1000;
6896 if (target_mw < port->operating_snk_mw) {
6897 ret = -EINVAL;
6898 goto port_unlock;
6899 }
6900
6901 port->upcoming_state = SNK_NEGOTIATE_PPS_CAPABILITIES;
6902 ret = tcpm_ams_start(port, POWER_NEGOTIATION);
6903 if (ret == -EAGAIN) {
6904 port->upcoming_state = INVALID_STATE;
6905 goto port_unlock;
6906 }
6907
6908 /* Round down output voltage to align with PPS valid steps */
6909 req_out_volt = req_out_volt - (req_out_volt % RDO_PROG_VOLT_MV_STEP);
6910
6911 reinit_completion(&port->pps_complete);
6912 port->pps_data.req_out_volt = req_out_volt;
6913 port->pps_status = 0;
6914 port->pps_pending = true;
6915 mutex_unlock(&port->lock);
6916
6917 if (!wait_for_completion_timeout(&port->pps_complete,
6918 msecs_to_jiffies(PD_PPS_CTRL_TIMEOUT)))
6919 ret = -ETIMEDOUT;
6920 else
6921 ret = port->pps_status;
6922
6923 goto swap_unlock;
6924
6925 port_unlock:
6926 mutex_unlock(&port->lock);
6927 swap_unlock:
6928 mutex_unlock(&port->swap_lock);
6929
6930 return ret;
6931 }
6932
tcpm_pps_activate(struct tcpm_port * port,bool activate)6933 static int tcpm_pps_activate(struct tcpm_port *port, bool activate)
6934 {
6935 int ret = 0;
6936
6937 mutex_lock(&port->swap_lock);
6938 mutex_lock(&port->lock);
6939
6940 if (!port->pps_data.supported) {
6941 ret = -EOPNOTSUPP;
6942 goto port_unlock;
6943 }
6944
6945 /* Trying to deactivate PPS when already deactivated so just bail */
6946 if (!port->pps_data.active && !activate)
6947 goto port_unlock;
6948
6949 if (port->state != SNK_READY) {
6950 ret = -EAGAIN;
6951 goto port_unlock;
6952 }
6953
6954 if (activate)
6955 port->upcoming_state = SNK_NEGOTIATE_PPS_CAPABILITIES;
6956 else
6957 port->upcoming_state = SNK_NEGOTIATE_CAPABILITIES;
6958 ret = tcpm_ams_start(port, POWER_NEGOTIATION);
6959 if (ret == -EAGAIN) {
6960 port->upcoming_state = INVALID_STATE;
6961 goto port_unlock;
6962 }
6963
6964 reinit_completion(&port->pps_complete);
6965 port->pps_status = 0;
6966 port->pps_pending = true;
6967
6968 /* Trigger PPS request or move back to standard PDO contract */
6969 if (activate) {
6970 port->pps_data.req_out_volt = port->supply_voltage;
6971 port->pps_data.req_op_curr = port->current_limit;
6972 }
6973 mutex_unlock(&port->lock);
6974
6975 if (!wait_for_completion_timeout(&port->pps_complete,
6976 msecs_to_jiffies(PD_PPS_CTRL_TIMEOUT)))
6977 ret = -ETIMEDOUT;
6978 else
6979 ret = port->pps_status;
6980
6981 goto swap_unlock;
6982
6983 port_unlock:
6984 mutex_unlock(&port->lock);
6985 swap_unlock:
6986 mutex_unlock(&port->swap_lock);
6987
6988 return ret;
6989 }
6990
tcpm_init(struct tcpm_port * port)6991 static void tcpm_init(struct tcpm_port *port)
6992 {
6993 enum typec_cc_status cc1, cc2;
6994
6995 port->tcpc->init(port->tcpc);
6996
6997 tcpm_reset_port(port);
6998
6999 /*
7000 * XXX
7001 * Should possibly wait for VBUS to settle if it was enabled locally
7002 * since tcpm_reset_port() will disable VBUS.
7003 */
7004 port->vbus_present = port->tcpc->get_vbus(port->tcpc);
7005 if (port->vbus_present)
7006 port->vbus_never_low = true;
7007
7008 /*
7009 * 1. When vbus_present is true, voltage on VBUS is already at VSAFE5V.
7010 * So implicitly vbus_vsafe0v = false.
7011 *
7012 * 2. When vbus_present is false and TCPC does NOT support querying
7013 * vsafe0v status, then, it's best to assume vbus is at VSAFE0V i.e.
7014 * vbus_vsafe0v is true.
7015 *
7016 * 3. When vbus_present is false and TCPC does support querying vsafe0v,
7017 * then, query tcpc for vsafe0v status.
7018 */
7019 if (port->vbus_present)
7020 port->vbus_vsafe0v = false;
7021 else if (!port->tcpc->is_vbus_vsafe0v)
7022 port->vbus_vsafe0v = true;
7023 else
7024 port->vbus_vsafe0v = port->tcpc->is_vbus_vsafe0v(port->tcpc);
7025
7026 tcpm_set_state(port, tcpm_default_state(port), 0);
7027
7028 if (port->tcpc->get_cc(port->tcpc, &cc1, &cc2) == 0)
7029 _tcpm_cc_change(port, cc1, cc2);
7030
7031 /*
7032 * Some adapters need a clean slate at startup, and won't recover
7033 * otherwise. So do not try to be fancy and force a clean disconnect.
7034 */
7035 tcpm_set_state(port, PORT_RESET, 0);
7036 }
7037
tcpm_port_type_set(struct typec_port * p,enum typec_port_type type)7038 static int tcpm_port_type_set(struct typec_port *p, enum typec_port_type type)
7039 {
7040 struct tcpm_port *port = typec_get_drvdata(p);
7041
7042 mutex_lock(&port->lock);
7043 if (type == port->port_type)
7044 goto port_unlock;
7045
7046 port->port_type = type;
7047
7048 if (!port->connected) {
7049 tcpm_set_state(port, PORT_RESET, 0);
7050 } else if (type == TYPEC_PORT_SNK) {
7051 if (!(port->pwr_role == TYPEC_SINK &&
7052 port->data_role == TYPEC_DEVICE))
7053 tcpm_set_state(port, PORT_RESET, 0);
7054 } else if (type == TYPEC_PORT_SRC) {
7055 if (!(port->pwr_role == TYPEC_SOURCE &&
7056 port->data_role == TYPEC_HOST))
7057 tcpm_set_state(port, PORT_RESET, 0);
7058 }
7059
7060 port_unlock:
7061 mutex_unlock(&port->lock);
7062 return 0;
7063 }
7064
tcpm_find_pd_data(struct tcpm_port * port,struct usb_power_delivery * pd)7065 static struct pd_data *tcpm_find_pd_data(struct tcpm_port *port, struct usb_power_delivery *pd)
7066 {
7067 int i;
7068
7069 for (i = 0; port->pd_list[i]; i++) {
7070 if (port->pd_list[i]->pd == pd)
7071 return port->pd_list[i];
7072 }
7073
7074 return ERR_PTR(-ENODATA);
7075 }
7076
tcpm_pd_get(struct typec_port * p)7077 static struct usb_power_delivery **tcpm_pd_get(struct typec_port *p)
7078 {
7079 struct tcpm_port *port = typec_get_drvdata(p);
7080
7081 return port->pds;
7082 }
7083
tcpm_pd_set(struct typec_port * p,struct usb_power_delivery * pd)7084 static int tcpm_pd_set(struct typec_port *p, struct usb_power_delivery *pd)
7085 {
7086 struct tcpm_port *port = typec_get_drvdata(p);
7087 struct pd_data *data;
7088 int i, ret = 0;
7089
7090 mutex_lock(&port->lock);
7091
7092 if (port->selected_pd == pd)
7093 goto unlock;
7094
7095 data = tcpm_find_pd_data(port, pd);
7096 if (IS_ERR(data)) {
7097 ret = PTR_ERR(data);
7098 goto unlock;
7099 }
7100
7101 if (data->sink_desc.pdo[0]) {
7102 for (i = 0; i < PDO_MAX_OBJECTS && data->sink_desc.pdo[i]; i++)
7103 port->snk_pdo[i] = data->sink_desc.pdo[i];
7104 port->nr_snk_pdo = i;
7105 port->operating_snk_mw = data->operating_snk_mw;
7106 }
7107
7108 if (data->source_desc.pdo[0]) {
7109 for (i = 0; i < PDO_MAX_OBJECTS && data->source_desc.pdo[i]; i++)
7110 port->src_pdo[i] = data->source_desc.pdo[i];
7111 port->nr_src_pdo = i;
7112 }
7113
7114 switch (port->state) {
7115 case SRC_UNATTACHED:
7116 case SRC_ATTACH_WAIT:
7117 case SRC_TRYWAIT:
7118 tcpm_set_cc(port, tcpm_rp_cc(port));
7119 break;
7120 case SRC_SEND_CAPABILITIES:
7121 case SRC_SEND_CAPABILITIES_TIMEOUT:
7122 case SRC_NEGOTIATE_CAPABILITIES:
7123 case SRC_READY:
7124 case SRC_WAIT_NEW_CAPABILITIES:
7125 port->caps_count = 0;
7126 port->upcoming_state = SRC_SEND_CAPABILITIES;
7127 ret = tcpm_ams_start(port, POWER_NEGOTIATION);
7128 if (ret == -EAGAIN) {
7129 port->upcoming_state = INVALID_STATE;
7130 goto unlock;
7131 }
7132 break;
7133 case SNK_NEGOTIATE_CAPABILITIES:
7134 case SNK_NEGOTIATE_PPS_CAPABILITIES:
7135 case SNK_READY:
7136 case SNK_TRANSITION_SINK:
7137 case SNK_TRANSITION_SINK_VBUS:
7138 if (port->pps_data.active)
7139 port->upcoming_state = SNK_NEGOTIATE_PPS_CAPABILITIES;
7140 else if (port->pd_capable)
7141 port->upcoming_state = SNK_NEGOTIATE_CAPABILITIES;
7142 else
7143 break;
7144
7145 port->update_sink_caps = true;
7146
7147 ret = tcpm_ams_start(port, POWER_NEGOTIATION);
7148 if (ret == -EAGAIN) {
7149 port->upcoming_state = INVALID_STATE;
7150 goto unlock;
7151 }
7152 break;
7153 default:
7154 break;
7155 }
7156
7157 port->port_source_caps = data->source_cap;
7158 port->port_sink_caps = data->sink_cap;
7159 typec_port_set_usb_power_delivery(p, NULL);
7160 port->selected_pd = pd;
7161 typec_port_set_usb_power_delivery(p, port->selected_pd);
7162 unlock:
7163 mutex_unlock(&port->lock);
7164 return ret;
7165 }
7166
7167 static const struct typec_operations tcpm_ops = {
7168 .try_role = tcpm_try_role,
7169 .dr_set = tcpm_dr_set,
7170 .pr_set = tcpm_pr_set,
7171 .vconn_set = tcpm_vconn_set,
7172 .port_type_set = tcpm_port_type_set,
7173 .pd_get = tcpm_pd_get,
7174 .pd_set = tcpm_pd_set
7175 };
7176
tcpm_tcpc_reset(struct tcpm_port * port)7177 void tcpm_tcpc_reset(struct tcpm_port *port)
7178 {
7179 mutex_lock(&port->lock);
7180 /* XXX: Maintain PD connection if possible? */
7181 tcpm_init(port);
7182 mutex_unlock(&port->lock);
7183 }
7184 EXPORT_SYMBOL_GPL(tcpm_tcpc_reset);
7185
tcpm_port_unregister_pd(struct tcpm_port * port)7186 static void tcpm_port_unregister_pd(struct tcpm_port *port)
7187 {
7188 int i;
7189
7190 port->port_sink_caps = NULL;
7191 port->port_source_caps = NULL;
7192 for (i = 0; i < port->pd_count; i++) {
7193 usb_power_delivery_unregister_capabilities(port->pd_list[i]->sink_cap);
7194 usb_power_delivery_unregister_capabilities(port->pd_list[i]->source_cap);
7195 devm_kfree(port->dev, port->pd_list[i]);
7196 port->pd_list[i] = NULL;
7197 usb_power_delivery_unregister(port->pds[i]);
7198 port->pds[i] = NULL;
7199 }
7200 }
7201
tcpm_port_register_pd(struct tcpm_port * port)7202 static int tcpm_port_register_pd(struct tcpm_port *port)
7203 {
7204 u16 pd_revision = port->typec_caps.pd_revision;
7205 u16 pd_version = port->pd_rev.ver_major << 8 | port->pd_rev.ver_minor;
7206 struct usb_power_delivery_desc desc = { pd_revision, pd_version };
7207 struct usb_power_delivery_capabilities *cap;
7208 int ret, i;
7209
7210 if (!port->nr_src_pdo && !port->nr_snk_pdo)
7211 return 0;
7212
7213 for (i = 0; i < port->pd_count; i++) {
7214 port->pds[i] = usb_power_delivery_register(port->dev, &desc);
7215 if (IS_ERR(port->pds[i])) {
7216 ret = PTR_ERR(port->pds[i]);
7217 goto err_unregister;
7218 }
7219 port->pd_list[i]->pd = port->pds[i];
7220
7221 if (port->pd_list[i]->source_desc.pdo[0]) {
7222 cap = usb_power_delivery_register_capabilities(port->pds[i],
7223 &port->pd_list[i]->source_desc);
7224 if (IS_ERR(cap)) {
7225 ret = PTR_ERR(cap);
7226 goto err_unregister;
7227 }
7228 port->pd_list[i]->source_cap = cap;
7229 }
7230
7231 if (port->pd_list[i]->sink_desc.pdo[0]) {
7232 cap = usb_power_delivery_register_capabilities(port->pds[i],
7233 &port->pd_list[i]->sink_desc);
7234 if (IS_ERR(cap)) {
7235 ret = PTR_ERR(cap);
7236 goto err_unregister;
7237 }
7238 port->pd_list[i]->sink_cap = cap;
7239 }
7240 }
7241
7242 port->port_source_caps = port->pd_list[0]->source_cap;
7243 port->port_sink_caps = port->pd_list[0]->sink_cap;
7244 port->selected_pd = port->pds[0];
7245 return 0;
7246
7247 err_unregister:
7248 tcpm_port_unregister_pd(port);
7249
7250 return ret;
7251 }
7252
tcpm_fw_get_timings(struct tcpm_port * port,struct fwnode_handle * fwnode)7253 static void tcpm_fw_get_timings(struct tcpm_port *port, struct fwnode_handle *fwnode)
7254 {
7255 int ret;
7256 u32 val;
7257
7258 ret = fwnode_property_read_u32(fwnode, "sink-wait-cap-time-ms", &val);
7259 if (!ret)
7260 port->timings.sink_wait_cap_time = val;
7261 else
7262 port->timings.sink_wait_cap_time = PD_T_SINK_WAIT_CAP;
7263
7264 ret = fwnode_property_read_u32(fwnode, "ps-source-off-time-ms", &val);
7265 if (!ret)
7266 port->timings.ps_src_off_time = val;
7267 else
7268 port->timings.ps_src_off_time = PD_T_PS_SOURCE_OFF;
7269
7270 ret = fwnode_property_read_u32(fwnode, "cc-debounce-time-ms", &val);
7271 if (!ret)
7272 port->timings.cc_debounce_time = val;
7273 else
7274 port->timings.cc_debounce_time = PD_T_CC_DEBOUNCE;
7275
7276 ret = fwnode_property_read_u32(fwnode, "sink-bc12-completion-time-ms", &val);
7277 if (!ret)
7278 port->timings.snk_bc12_cmpletion_time = val;
7279 }
7280
tcpm_fw_get_caps(struct tcpm_port * port,struct fwnode_handle * fwnode)7281 static int tcpm_fw_get_caps(struct tcpm_port *port, struct fwnode_handle *fwnode)
7282 {
7283 struct fwnode_handle *capabilities, *caps = NULL;
7284 unsigned int nr_src_pdo, nr_snk_pdo;
7285 const char *opmode_str;
7286 u32 *src_pdo, *snk_pdo;
7287 u32 uw, frs_current;
7288 int ret = 0, i;
7289 int mode;
7290
7291 if (!fwnode)
7292 return -EINVAL;
7293
7294 /*
7295 * This fwnode has a "compatible" property, but is never populated as a
7296 * struct device. Instead we simply parse it to read the properties.
7297 * This it breaks fw_devlink=on. To maintain backward compatibility
7298 * with existing DT files, we work around this by deleting any
7299 * fwnode_links to/from this fwnode.
7300 */
7301 fw_devlink_purge_absent_suppliers(fwnode);
7302
7303 ret = typec_get_fw_cap(&port->typec_caps, fwnode);
7304 if (ret < 0)
7305 return ret;
7306
7307 mode = 0;
7308
7309 if (fwnode_property_read_bool(fwnode, "accessory-mode-audio"))
7310 port->typec_caps.accessory[mode++] = TYPEC_ACCESSORY_AUDIO;
7311
7312 if (fwnode_property_read_bool(fwnode, "accessory-mode-debug"))
7313 port->typec_caps.accessory[mode++] = TYPEC_ACCESSORY_DEBUG;
7314
7315 port->port_type = port->typec_caps.type;
7316 port->pd_supported = !fwnode_property_read_bool(fwnode, "pd-disable");
7317 port->slow_charger_loop = fwnode_property_read_bool(fwnode, "slow-charger-loop");
7318 port->self_powered = fwnode_property_read_bool(fwnode, "self-powered");
7319
7320 if (!port->pd_supported) {
7321 ret = fwnode_property_read_string(fwnode, "typec-power-opmode", &opmode_str);
7322 if (ret)
7323 return ret;
7324 ret = typec_find_pwr_opmode(opmode_str);
7325 if (ret < 0)
7326 return ret;
7327 port->src_rp = tcpm_pwr_opmode_to_rp(ret);
7328 return 0;
7329 }
7330
7331 /* The following code are applicable to pd-capable ports, i.e. pd_supported is true. */
7332
7333 /* FRS can only be supported by DRP ports */
7334 if (port->port_type == TYPEC_PORT_DRP) {
7335 ret = fwnode_property_read_u32(fwnode, "new-source-frs-typec-current",
7336 &frs_current);
7337 if (!ret && frs_current <= FRS_5V_3A)
7338 port->new_source_frs_current = frs_current;
7339
7340 if (ret)
7341 ret = 0;
7342 }
7343
7344 /* For the backward compatibility, "capabilities" node is optional. */
7345 capabilities = fwnode_get_named_child_node(fwnode, "capabilities");
7346 if (!capabilities) {
7347 port->pd_count = 1;
7348 } else {
7349 port->pd_count = fwnode_get_child_node_count(capabilities);
7350 if (!port->pd_count) {
7351 ret = -ENODATA;
7352 goto put_capabilities;
7353 }
7354 }
7355
7356 port->pds = devm_kcalloc(port->dev, port->pd_count, sizeof(struct usb_power_delivery *),
7357 GFP_KERNEL);
7358 if (!port->pds) {
7359 ret = -ENOMEM;
7360 goto put_capabilities;
7361 }
7362
7363 port->pd_list = devm_kcalloc(port->dev, port->pd_count, sizeof(struct pd_data *),
7364 GFP_KERNEL);
7365 if (!port->pd_list) {
7366 ret = -ENOMEM;
7367 goto put_capabilities;
7368 }
7369
7370 for (i = 0; i < port->pd_count; i++) {
7371 port->pd_list[i] = devm_kzalloc(port->dev, sizeof(struct pd_data), GFP_KERNEL);
7372 if (!port->pd_list[i]) {
7373 ret = -ENOMEM;
7374 goto put_capabilities;
7375 }
7376
7377 src_pdo = port->pd_list[i]->source_desc.pdo;
7378 port->pd_list[i]->source_desc.role = TYPEC_SOURCE;
7379 snk_pdo = port->pd_list[i]->sink_desc.pdo;
7380 port->pd_list[i]->sink_desc.role = TYPEC_SINK;
7381
7382 /* If "capabilities" is NULL, fall back to single pd cap population. */
7383 if (!capabilities)
7384 caps = fwnode;
7385 else
7386 caps = fwnode_get_next_child_node(capabilities, caps);
7387
7388 if (port->port_type != TYPEC_PORT_SNK) {
7389 ret = fwnode_property_count_u32(caps, "source-pdos");
7390 if (ret == 0) {
7391 ret = -EINVAL;
7392 goto put_caps;
7393 }
7394 if (ret < 0)
7395 goto put_caps;
7396
7397 nr_src_pdo = min(ret, PDO_MAX_OBJECTS);
7398 ret = fwnode_property_read_u32_array(caps, "source-pdos", src_pdo,
7399 nr_src_pdo);
7400 if (ret)
7401 goto put_caps;
7402
7403 ret = tcpm_validate_caps(port, src_pdo, nr_src_pdo);
7404 if (ret)
7405 goto put_caps;
7406
7407 if (i == 0) {
7408 port->nr_src_pdo = nr_src_pdo;
7409 memcpy_and_pad(port->src_pdo, sizeof(u32) * PDO_MAX_OBJECTS,
7410 port->pd_list[0]->source_desc.pdo,
7411 sizeof(u32) * nr_src_pdo,
7412 0);
7413 }
7414 }
7415
7416 if (port->port_type != TYPEC_PORT_SRC) {
7417 ret = fwnode_property_count_u32(caps, "sink-pdos");
7418 if (ret == 0) {
7419 ret = -EINVAL;
7420 goto put_caps;
7421 }
7422
7423 if (ret < 0)
7424 goto put_caps;
7425
7426 nr_snk_pdo = min(ret, PDO_MAX_OBJECTS);
7427 ret = fwnode_property_read_u32_array(caps, "sink-pdos", snk_pdo,
7428 nr_snk_pdo);
7429 if (ret)
7430 goto put_caps;
7431
7432 ret = tcpm_validate_caps(port, snk_pdo, nr_snk_pdo);
7433 if (ret)
7434 goto put_caps;
7435
7436 if (fwnode_property_read_u32(caps, "op-sink-microwatt", &uw) < 0) {
7437 ret = -EINVAL;
7438 goto put_caps;
7439 }
7440
7441 port->pd_list[i]->operating_snk_mw = uw / 1000;
7442
7443 if (i == 0) {
7444 port->nr_snk_pdo = nr_snk_pdo;
7445 memcpy_and_pad(port->snk_pdo, sizeof(u32) * PDO_MAX_OBJECTS,
7446 port->pd_list[0]->sink_desc.pdo,
7447 sizeof(u32) * nr_snk_pdo,
7448 0);
7449 port->operating_snk_mw = port->pd_list[0]->operating_snk_mw;
7450 }
7451 }
7452 }
7453
7454 put_caps:
7455 if (caps != fwnode)
7456 fwnode_handle_put(caps);
7457 put_capabilities:
7458 fwnode_handle_put(capabilities);
7459 return ret;
7460 }
7461
tcpm_fw_get_snk_vdos(struct tcpm_port * port,struct fwnode_handle * fwnode)7462 static int tcpm_fw_get_snk_vdos(struct tcpm_port *port, struct fwnode_handle *fwnode)
7463 {
7464 int ret;
7465
7466 /* sink-vdos is optional */
7467 ret = fwnode_property_count_u32(fwnode, "sink-vdos");
7468 if (ret < 0)
7469 return 0;
7470
7471 port->nr_snk_vdo = min(ret, VDO_MAX_OBJECTS);
7472 if (port->nr_snk_vdo) {
7473 ret = fwnode_property_read_u32_array(fwnode, "sink-vdos",
7474 port->snk_vdo,
7475 port->nr_snk_vdo);
7476 if (ret < 0)
7477 return ret;
7478 }
7479
7480 /* If sink-vdos is found, sink-vdos-v1 is expected for backward compatibility. */
7481 if (port->nr_snk_vdo) {
7482 ret = fwnode_property_count_u32(fwnode, "sink-vdos-v1");
7483 if (ret < 0)
7484 return ret;
7485 else if (ret == 0)
7486 return -ENODATA;
7487
7488 port->nr_snk_vdo_v1 = min(ret, VDO_MAX_OBJECTS);
7489 ret = fwnode_property_read_u32_array(fwnode, "sink-vdos-v1",
7490 port->snk_vdo_v1,
7491 port->nr_snk_vdo_v1);
7492 if (ret < 0)
7493 return ret;
7494 }
7495
7496 return 0;
7497 }
7498
tcpm_fw_get_pd_revision(struct tcpm_port * port,struct fwnode_handle * fwnode)7499 static void tcpm_fw_get_pd_revision(struct tcpm_port *port, struct fwnode_handle *fwnode)
7500 {
7501 int ret;
7502 u8 val[4];
7503
7504 ret = fwnode_property_count_u8(fwnode, "pd-revision");
7505 if (!ret || ret != 4) {
7506 tcpm_log(port, "Unable to find pd-revision property or incorrect array size");
7507 return;
7508 }
7509
7510 ret = fwnode_property_read_u8_array(fwnode, "pd-revision", val, 4);
7511 if (ret) {
7512 tcpm_log(port, "Failed to parse pd-revision, ret:(%d)", ret);
7513 return;
7514 }
7515
7516 port->pd_rev.rev_major = val[0];
7517 port->pd_rev.rev_minor = val[1];
7518 port->pd_rev.ver_major = val[2];
7519 port->pd_rev.ver_minor = val[3];
7520 }
7521
7522 /* Power Supply access to expose source power information */
7523 enum tcpm_psy_online_states {
7524 TCPM_PSY_OFFLINE = 0,
7525 TCPM_PSY_FIXED_ONLINE,
7526 TCPM_PSY_PROG_ONLINE,
7527 };
7528
7529 static enum power_supply_property tcpm_psy_props[] = {
7530 POWER_SUPPLY_PROP_USB_TYPE,
7531 POWER_SUPPLY_PROP_ONLINE,
7532 POWER_SUPPLY_PROP_VOLTAGE_MIN,
7533 POWER_SUPPLY_PROP_VOLTAGE_MAX,
7534 POWER_SUPPLY_PROP_VOLTAGE_NOW,
7535 POWER_SUPPLY_PROP_CURRENT_MAX,
7536 POWER_SUPPLY_PROP_CURRENT_NOW,
7537 };
7538
tcpm_psy_get_online(struct tcpm_port * port,union power_supply_propval * val)7539 static int tcpm_psy_get_online(struct tcpm_port *port,
7540 union power_supply_propval *val)
7541 {
7542 if (port->vbus_charge) {
7543 if (port->pps_data.active)
7544 val->intval = TCPM_PSY_PROG_ONLINE;
7545 else
7546 val->intval = TCPM_PSY_FIXED_ONLINE;
7547 } else {
7548 val->intval = TCPM_PSY_OFFLINE;
7549 }
7550
7551 return 0;
7552 }
7553
tcpm_psy_get_voltage_min(struct tcpm_port * port,union power_supply_propval * val)7554 static int tcpm_psy_get_voltage_min(struct tcpm_port *port,
7555 union power_supply_propval *val)
7556 {
7557 if (port->pps_data.active)
7558 val->intval = port->pps_data.min_volt * 1000;
7559 else
7560 val->intval = port->supply_voltage * 1000;
7561
7562 return 0;
7563 }
7564
tcpm_psy_get_voltage_max(struct tcpm_port * port,union power_supply_propval * val)7565 static int tcpm_psy_get_voltage_max(struct tcpm_port *port,
7566 union power_supply_propval *val)
7567 {
7568 if (port->pps_data.active)
7569 val->intval = port->pps_data.max_volt * 1000;
7570 else
7571 val->intval = port->supply_voltage * 1000;
7572
7573 return 0;
7574 }
7575
tcpm_psy_get_voltage_now(struct tcpm_port * port,union power_supply_propval * val)7576 static int tcpm_psy_get_voltage_now(struct tcpm_port *port,
7577 union power_supply_propval *val)
7578 {
7579 val->intval = port->supply_voltage * 1000;
7580
7581 return 0;
7582 }
7583
tcpm_psy_get_current_max(struct tcpm_port * port,union power_supply_propval * val)7584 static int tcpm_psy_get_current_max(struct tcpm_port *port,
7585 union power_supply_propval *val)
7586 {
7587 if (port->pps_data.active)
7588 val->intval = port->pps_data.max_curr * 1000;
7589 else
7590 val->intval = port->current_limit * 1000;
7591
7592 return 0;
7593 }
7594
tcpm_psy_get_current_now(struct tcpm_port * port,union power_supply_propval * val)7595 static int tcpm_psy_get_current_now(struct tcpm_port *port,
7596 union power_supply_propval *val)
7597 {
7598 val->intval = port->current_limit * 1000;
7599
7600 return 0;
7601 }
7602
tcpm_psy_get_input_power_limit(struct tcpm_port * port,union power_supply_propval * val)7603 static int tcpm_psy_get_input_power_limit(struct tcpm_port *port,
7604 union power_supply_propval *val)
7605 {
7606 unsigned int src_mv, src_ma, max_src_uw = 0;
7607 unsigned int i, tmp;
7608
7609 for (i = 0; i < port->nr_source_caps; i++) {
7610 u32 pdo = port->source_caps[i];
7611
7612 if (pdo_type(pdo) == PDO_TYPE_FIXED) {
7613 src_mv = pdo_fixed_voltage(pdo);
7614 src_ma = pdo_max_current(pdo);
7615 tmp = src_mv * src_ma;
7616 max_src_uw = max(tmp, max_src_uw);
7617 }
7618 }
7619
7620 val->intval = max_src_uw;
7621 return 0;
7622 }
7623
tcpm_psy_get_prop(struct power_supply * psy,enum power_supply_property psp,union power_supply_propval * val)7624 static int tcpm_psy_get_prop(struct power_supply *psy,
7625 enum power_supply_property psp,
7626 union power_supply_propval *val)
7627 {
7628 struct tcpm_port *port = power_supply_get_drvdata(psy);
7629 int ret = 0;
7630
7631 switch (psp) {
7632 case POWER_SUPPLY_PROP_USB_TYPE:
7633 val->intval = port->usb_type;
7634 break;
7635 case POWER_SUPPLY_PROP_ONLINE:
7636 ret = tcpm_psy_get_online(port, val);
7637 break;
7638 case POWER_SUPPLY_PROP_VOLTAGE_MIN:
7639 ret = tcpm_psy_get_voltage_min(port, val);
7640 break;
7641 case POWER_SUPPLY_PROP_VOLTAGE_MAX:
7642 ret = tcpm_psy_get_voltage_max(port, val);
7643 break;
7644 case POWER_SUPPLY_PROP_VOLTAGE_NOW:
7645 ret = tcpm_psy_get_voltage_now(port, val);
7646 break;
7647 case POWER_SUPPLY_PROP_CURRENT_MAX:
7648 ret = tcpm_psy_get_current_max(port, val);
7649 break;
7650 case POWER_SUPPLY_PROP_CURRENT_NOW:
7651 ret = tcpm_psy_get_current_now(port, val);
7652 break;
7653 case POWER_SUPPLY_PROP_INPUT_POWER_LIMIT:
7654 tcpm_psy_get_input_power_limit(port, val);
7655 break;
7656 default:
7657 ret = -EINVAL;
7658 break;
7659 }
7660
7661 return ret;
7662 }
7663
tcpm_psy_set_online(struct tcpm_port * port,const union power_supply_propval * val)7664 static int tcpm_psy_set_online(struct tcpm_port *port,
7665 const union power_supply_propval *val)
7666 {
7667 int ret;
7668
7669 switch (val->intval) {
7670 case TCPM_PSY_FIXED_ONLINE:
7671 ret = tcpm_pps_activate(port, false);
7672 break;
7673 case TCPM_PSY_PROG_ONLINE:
7674 ret = tcpm_pps_activate(port, true);
7675 break;
7676 default:
7677 ret = -EINVAL;
7678 break;
7679 }
7680
7681 return ret;
7682 }
7683
tcpm_psy_set_prop(struct power_supply * psy,enum power_supply_property psp,const union power_supply_propval * val)7684 static int tcpm_psy_set_prop(struct power_supply *psy,
7685 enum power_supply_property psp,
7686 const union power_supply_propval *val)
7687 {
7688 struct tcpm_port *port = power_supply_get_drvdata(psy);
7689 int ret;
7690
7691 /*
7692 * All the properties below are related to USB PD. The check needs to be
7693 * property specific when a non-pd related property is added.
7694 */
7695 if (!port->pd_supported)
7696 return -EOPNOTSUPP;
7697
7698 switch (psp) {
7699 case POWER_SUPPLY_PROP_ONLINE:
7700 ret = tcpm_psy_set_online(port, val);
7701 break;
7702 case POWER_SUPPLY_PROP_VOLTAGE_NOW:
7703 ret = tcpm_pps_set_out_volt(port, val->intval / 1000);
7704 break;
7705 case POWER_SUPPLY_PROP_CURRENT_NOW:
7706 if (val->intval > port->pps_data.max_curr * 1000)
7707 ret = -EINVAL;
7708 else
7709 ret = tcpm_pps_set_op_curr(port, val->intval / 1000);
7710 break;
7711 default:
7712 ret = -EINVAL;
7713 break;
7714 }
7715 power_supply_changed(port->psy);
7716 return ret;
7717 }
7718
tcpm_psy_prop_writeable(struct power_supply * psy,enum power_supply_property psp)7719 static int tcpm_psy_prop_writeable(struct power_supply *psy,
7720 enum power_supply_property psp)
7721 {
7722 switch (psp) {
7723 case POWER_SUPPLY_PROP_ONLINE:
7724 case POWER_SUPPLY_PROP_VOLTAGE_NOW:
7725 case POWER_SUPPLY_PROP_CURRENT_NOW:
7726 return 1;
7727 default:
7728 return 0;
7729 }
7730 }
7731
7732 static const char *tcpm_psy_name_prefix = "tcpm-source-psy-";
7733
devm_tcpm_psy_register(struct tcpm_port * port)7734 static int devm_tcpm_psy_register(struct tcpm_port *port)
7735 {
7736 struct power_supply_config psy_cfg = {};
7737 const char *port_dev_name = dev_name(port->dev);
7738 size_t psy_name_len = strlen(tcpm_psy_name_prefix) +
7739 strlen(port_dev_name) + 1;
7740 char *psy_name;
7741
7742 psy_cfg.drv_data = port;
7743 psy_cfg.fwnode = dev_fwnode(port->dev);
7744 psy_name = devm_kzalloc(port->dev, psy_name_len, GFP_KERNEL);
7745 if (!psy_name)
7746 return -ENOMEM;
7747
7748 snprintf(psy_name, psy_name_len, "%s%s", tcpm_psy_name_prefix,
7749 port_dev_name);
7750 port->psy_desc.name = psy_name;
7751 port->psy_desc.type = POWER_SUPPLY_TYPE_USB;
7752 port->psy_desc.usb_types = BIT(POWER_SUPPLY_USB_TYPE_C) |
7753 BIT(POWER_SUPPLY_USB_TYPE_PD) |
7754 BIT(POWER_SUPPLY_USB_TYPE_PD_PPS);
7755 port->psy_desc.properties = tcpm_psy_props;
7756 port->psy_desc.num_properties = ARRAY_SIZE(tcpm_psy_props);
7757 port->psy_desc.get_property = tcpm_psy_get_prop;
7758 port->psy_desc.set_property = tcpm_psy_set_prop;
7759 port->psy_desc.property_is_writeable = tcpm_psy_prop_writeable;
7760
7761 port->usb_type = POWER_SUPPLY_USB_TYPE_C;
7762
7763 port->psy = devm_power_supply_register(port->dev, &port->psy_desc,
7764 &psy_cfg);
7765
7766 return PTR_ERR_OR_ZERO(port->psy);
7767 }
7768
state_machine_timer_handler(struct hrtimer * timer)7769 static enum hrtimer_restart state_machine_timer_handler(struct hrtimer *timer)
7770 {
7771 struct tcpm_port *port = container_of(timer, struct tcpm_port, state_machine_timer);
7772
7773 if (port->registered)
7774 kthread_queue_work(port->wq, &port->state_machine);
7775 return HRTIMER_NORESTART;
7776 }
7777
vdm_state_machine_timer_handler(struct hrtimer * timer)7778 static enum hrtimer_restart vdm_state_machine_timer_handler(struct hrtimer *timer)
7779 {
7780 struct tcpm_port *port = container_of(timer, struct tcpm_port, vdm_state_machine_timer);
7781
7782 if (port->registered)
7783 kthread_queue_work(port->wq, &port->vdm_state_machine);
7784 return HRTIMER_NORESTART;
7785 }
7786
enable_frs_timer_handler(struct hrtimer * timer)7787 static enum hrtimer_restart enable_frs_timer_handler(struct hrtimer *timer)
7788 {
7789 struct tcpm_port *port = container_of(timer, struct tcpm_port, enable_frs_timer);
7790
7791 if (port->registered)
7792 kthread_queue_work(port->wq, &port->enable_frs);
7793 return HRTIMER_NORESTART;
7794 }
7795
send_discover_timer_handler(struct hrtimer * timer)7796 static enum hrtimer_restart send_discover_timer_handler(struct hrtimer *timer)
7797 {
7798 struct tcpm_port *port = container_of(timer, struct tcpm_port, send_discover_timer);
7799
7800 if (port->registered)
7801 kthread_queue_work(port->wq, &port->send_discover_work);
7802 return HRTIMER_NORESTART;
7803 }
7804
tcpm_register_port(struct device * dev,struct tcpc_dev * tcpc)7805 struct tcpm_port *tcpm_register_port(struct device *dev, struct tcpc_dev *tcpc)
7806 {
7807 struct tcpm_port *port;
7808 int err;
7809
7810 if (!dev || !tcpc ||
7811 !tcpc->get_vbus || !tcpc->set_cc || !tcpc->get_cc ||
7812 !tcpc->set_polarity || !tcpc->set_vconn || !tcpc->set_vbus ||
7813 !tcpc->set_pd_rx || !tcpc->set_roles || !tcpc->pd_transmit)
7814 return ERR_PTR(-EINVAL);
7815
7816 port = devm_kzalloc(dev, sizeof(*port), GFP_KERNEL);
7817 if (!port)
7818 return ERR_PTR(-ENOMEM);
7819
7820 port->dev = dev;
7821 port->tcpc = tcpc;
7822
7823 mutex_init(&port->lock);
7824 mutex_init(&port->swap_lock);
7825
7826 port->wq = kthread_run_worker(0, dev_name(dev));
7827 if (IS_ERR(port->wq))
7828 return ERR_CAST(port->wq);
7829 sched_set_fifo(port->wq->task);
7830
7831 kthread_init_work(&port->state_machine, tcpm_state_machine_work);
7832 kthread_init_work(&port->vdm_state_machine, vdm_state_machine_work);
7833 kthread_init_work(&port->event_work, tcpm_pd_event_handler);
7834 kthread_init_work(&port->enable_frs, tcpm_enable_frs_work);
7835 kthread_init_work(&port->send_discover_work, tcpm_send_discover_work);
7836 hrtimer_setup(&port->state_machine_timer, state_machine_timer_handler, CLOCK_MONOTONIC,
7837 HRTIMER_MODE_REL);
7838 hrtimer_setup(&port->vdm_state_machine_timer, vdm_state_machine_timer_handler,
7839 CLOCK_MONOTONIC, HRTIMER_MODE_REL);
7840 hrtimer_setup(&port->enable_frs_timer, enable_frs_timer_handler, CLOCK_MONOTONIC,
7841 HRTIMER_MODE_REL);
7842 hrtimer_setup(&port->send_discover_timer, send_discover_timer_handler, CLOCK_MONOTONIC,
7843 HRTIMER_MODE_REL);
7844
7845 spin_lock_init(&port->pd_event_lock);
7846
7847 init_completion(&port->tx_complete);
7848 init_completion(&port->swap_complete);
7849 init_completion(&port->pps_complete);
7850 tcpm_debugfs_init(port);
7851
7852 err = tcpm_fw_get_caps(port, tcpc->fwnode);
7853 if (err < 0)
7854 goto out_destroy_wq;
7855 err = tcpm_fw_get_snk_vdos(port, tcpc->fwnode);
7856 if (err < 0)
7857 goto out_destroy_wq;
7858
7859 tcpm_fw_get_timings(port, tcpc->fwnode);
7860 tcpm_fw_get_pd_revision(port, tcpc->fwnode);
7861
7862 port->try_role = port->typec_caps.prefer_role;
7863
7864 port->typec_caps.revision = 0x0120; /* Type-C spec release 1.2 */
7865
7866 if (port->pd_rev.rev_major)
7867 port->typec_caps.pd_revision = port->pd_rev.rev_major << 8 |
7868 port->pd_rev.rev_minor;
7869 else
7870 port->typec_caps.pd_revision = 0x0300; /* USB-PD spec release 3.0 */
7871
7872 port->typec_caps.svdm_version = SVDM_VER_2_0;
7873 port->typec_caps.driver_data = port;
7874 port->typec_caps.ops = &tcpm_ops;
7875 port->typec_caps.orientation_aware = 1;
7876
7877 port->partner_desc.identity = &port->partner_ident;
7878
7879 port->role_sw = usb_role_switch_get(port->dev);
7880 if (!port->role_sw)
7881 port->role_sw = fwnode_usb_role_switch_get(tcpc->fwnode);
7882 if (IS_ERR(port->role_sw)) {
7883 err = PTR_ERR(port->role_sw);
7884 goto out_destroy_wq;
7885 }
7886
7887 err = devm_tcpm_psy_register(port);
7888 if (err)
7889 goto out_role_sw_put;
7890 power_supply_changed(port->psy);
7891
7892 err = tcpm_port_register_pd(port);
7893 if (err)
7894 goto out_role_sw_put;
7895
7896 if (port->pds)
7897 port->typec_caps.pd = port->pds[0];
7898
7899 port->typec_port = typec_register_port(port->dev, &port->typec_caps);
7900 if (IS_ERR(port->typec_port)) {
7901 err = PTR_ERR(port->typec_port);
7902 goto out_unregister_pd;
7903 }
7904
7905 typec_port_register_altmodes(port->typec_port,
7906 &tcpm_altmode_ops, port,
7907 port->port_altmode, ALTMODE_DISCOVERY_MAX);
7908 typec_port_register_cable_ops(port->port_altmode, ARRAY_SIZE(port->port_altmode),
7909 &tcpm_cable_ops);
7910 port->registered = true;
7911
7912 mutex_lock(&port->lock);
7913 tcpm_init(port);
7914 mutex_unlock(&port->lock);
7915
7916 tcpm_log(port, "%s: registered", dev_name(dev));
7917 return port;
7918
7919 out_unregister_pd:
7920 tcpm_port_unregister_pd(port);
7921 out_role_sw_put:
7922 usb_role_switch_put(port->role_sw);
7923 out_destroy_wq:
7924 tcpm_debugfs_exit(port);
7925 kthread_destroy_worker(port->wq);
7926 return ERR_PTR(err);
7927 }
7928 EXPORT_SYMBOL_GPL(tcpm_register_port);
7929
tcpm_unregister_port(struct tcpm_port * port)7930 void tcpm_unregister_port(struct tcpm_port *port)
7931 {
7932 int i;
7933
7934 port->registered = false;
7935 kthread_destroy_worker(port->wq);
7936
7937 hrtimer_cancel(&port->send_discover_timer);
7938 hrtimer_cancel(&port->enable_frs_timer);
7939 hrtimer_cancel(&port->vdm_state_machine_timer);
7940 hrtimer_cancel(&port->state_machine_timer);
7941
7942 tcpm_reset_port(port);
7943
7944 tcpm_port_unregister_pd(port);
7945
7946 for (i = 0; i < ARRAY_SIZE(port->port_altmode); i++)
7947 typec_unregister_altmode(port->port_altmode[i]);
7948 typec_unregister_port(port->typec_port);
7949 usb_role_switch_put(port->role_sw);
7950 tcpm_debugfs_exit(port);
7951 }
7952 EXPORT_SYMBOL_GPL(tcpm_unregister_port);
7953
7954 MODULE_AUTHOR("Guenter Roeck <groeck@chromium.org>");
7955 MODULE_DESCRIPTION("USB Type-C Port Manager");
7956 MODULE_LICENSE("GPL");
7957