Home
last modified time | relevance | path

Searched full:recovery (Results 1 – 25 of 1356) sorted by relevance

12345678910>>...55

/linux/drivers/target/iscsi/
H A Discsi_target_datain_values.c76 * For Normal and Recovery DataSequenceInOrder=Yes and DataPDUInOrder=Yes.
90 if (dr->recovery && dr->generate_recovery_values) { in iscsit_set_datain_values_yes_and_yes()
98 next_burst_len = (!dr->recovery) ? in iscsit_set_datain_values_yes_and_yes()
100 read_data_done = (!dr->recovery) ? in iscsit_set_datain_values_yes_and_yes()
136 datain->data_sn = (!dr->recovery) ? cmd->data_sn++ : dr->data_sn++; in iscsit_set_datain_values_yes_and_yes()
139 if (!dr->recovery) { in iscsit_set_datain_values_yes_and_yes()
147 if (!dr->recovery) { in iscsit_set_datain_values_yes_and_yes()
157 (dr->recovery == DATAIN_WITHIN_COMMAND_RECOVERY) ? in iscsit_set_datain_values_yes_and_yes()
164 (dr->recovery == DATAIN_WITHIN_COMMAND_RECOVERY) ? in iscsit_set_datain_values_yes_and_yes()
174 * For Normal and Recovery DataSequenceInOrder=No and DataPDUInOrder=Yes.
[all …]
/linux/fs/ocfs2/dlm/
H A Ddlmrecovery.c5 * recovery stuff
127 /* Worker function used during recovery. */
164 * RECOVERY THREAD
169 /* wake the recovery thread in dlm_kick_recovery_thread()
171 * 1) sleeping with no recovery happening in dlm_kick_recovery_thread()
172 * 2) sleeping with recovery mastered elsewhere in dlm_kick_recovery_thread()
173 * 3) recovery mastered here, waiting on reco data */ in dlm_kick_recovery_thread()
178 /* Launch the recovery thread */
181 mlog(0, "starting dlm recovery thread...\n"); in dlm_launch_recovery_thread()
197 mlog(0, "waiting for dlm recovery thread to exit\n"); in dlm_complete_recovery_thread()
[all …]
/linux/drivers/s390/scsi/
H A Dzfcp_dbf.h28 * struct zfcp_dbf_rec_trigger - trace record for triggered recovery action
29 * @ready: number of ready recovery actions
30 * @running: number of running recovery actions
31 * @want: wanted recovery action
32 * @need: needed recovery action
42 * struct zfcp_dbf_rec_running - trace record for running recovery
45 * @rec_step: current step of the recovery action
58 * enum zfcp_dbf_rec_id - recovery trace record id
59 * @ZFCP_DBF_REC_TRIG: triggered recovery identifier
60 * @ZFCP_DBF_REC_RUN: running recovery identifier
[all …]
H A Dzfcp_def.h94 * @ZFCP_ERP_ACTION_REOPEN_LUN: LUN recovery.
95 * @ZFCP_ERP_ACTION_REOPEN_PORT: Port recovery.
96 * @ZFCP_ERP_ACTION_REOPEN_PORT_FORCED: Forced port recovery.
97 * @ZFCP_ERP_ACTION_REOPEN_ADAPTER: Adapter recovery.
130 u32 status; /* recovery status */
176 struct list_head erp_ready_head; /* error recovery for this
182 struct zfcp_erp_action erp_action; /* pending error recovery */
219 struct zfcp_erp_action erp_action; /* pending error recovery */
261 * running I/O and recovery is in struct zfcp_scsi_dev.
/linux/Documentation/PCI/
H A Dpci-error-recovery.rst4 PCI Error Recovery
26 of a bus disconnection, and then performing error recovery.
29 Reporting and recovery is performed in several steps. First, when
38 Next, recovery is performed in several stages. Most of the complexity
57 through a large number of contortions to complete recovery. Almost all
61 device drivers already handle very similar recovery procedures;
72 The error recovery API support is exposed to the driver in the form of
75 and the actual recovery steps taken are platform dependent. The
111 is assumed that the driver is not doing any direct recovery and requires
160 particular, if the platform doesn't isolate slots), and recovery
[all …]
/linux/drivers/md/dm-vdo/
H A Drecovery-journal.c6 #include "recovery-journal.h"
30 * The number of reserved blocks must be large enough to prevent a new recovery journal
40 * A lock_counter is intended to keep all of the locks for the blocks in the recovery journal. The
49 * Lock sets are laid out with the set for recovery journal first, followed by the logical zones,
61 * @journal: The recovery journal.
78 * @journal: The recovery journal.
109 * @journal: The recovery journal.
128 * vdo_release_recovery_journal_block_reference() - Release a reference to a recovery journal
130 * @journal: The recovery journal.
214 * is_block_dirty() - Check whether a recovery block is dirty.
[all …]
H A Dencodings.h197 /* The state of the recovery journal as encoded in the VDO super block. */
212 * A recovery journal entry stores three physical locations: a data location that is the value of a
225 /* The packed, on-disk representation of a recovery journal entry. */
266 /* The packed, on-disk representation of an old format recovery journal entry. */
322 * The packed, on-disk representation of a recovery journal block header. All fields are kept in
338 /* 8-bit metadata type (should always be one for the recovery journal) */
377 /* The number of entries in a v1 recovery journal block. */
391 /* The absolute position of an entry in a recovery journal or slab journal. */
477 /* Recovery journal point for last entry */
496 /* Recovery journal point for the last entry, packed into 64 bits */
[all …]
/linux/fs/gfs2/
H A Dlock_dlm.c20 #include "recovery.h"
365 * dlm/gfs2 recovery coordination using dlm_recover callbacks
371 * 4. dlm_controld starts and finishes its own user level recovery
372 * 5. dlm_controld starts dlm-kernel dlm_recoverd to do kernel recovery
374 * 7. dlm_recoverd does its own lock recovery
380 * 13. gfs2_recover provides recovery results to gfs2_control (recovery_result)
384 * - failures during recovery
411 * - parallel recovery steps across all nodes
426 * and missing a journal recovery?
437 * require recovery, because the mount in step 4 could not have
[all …]
/linux/drivers/remoteproc/
H A Dremoteproc_debugfs.c61 * disabled: By default coredump collection is disabled. Recovery will
68 * recovery process will have to wait until data is read by
174 /* expose recovery flag via debugfs */
185 * By writing to the 'recovery' debugfs entry, we control the behavior of the
186 * recovery mechanism dynamically. The default value of this entry is "enabled".
188 * The 'recovery' debugfs entry supports these commands:
192 * processor crashes while recovery is disabled, it will
193 * be automatically recovered too as soon as recovery is enabled.
199 * recover: This function will trigger an immediate recovery if the
201 * or checking the recovery state (enabled/disabled).
[all …]
/linux/Documentation/ABI/testing/
H A Dsysfs-class-remoteproc83 copy the dump. Also recovery process will not proceed until
86 What: /sys/class/remoteproc/.../recovery
89 Description: Remote processor recovery mechanism
91 Reports the recovery mechanism of the remote processor,
99 processor crashes while recovery is disabled, it will
100 be automatically recovered too as soon as recovery is enabled.
/linux/Documentation/networking/devlink/
H A Ddevlink-health.rst23 attributes of the health reporting and recovery procedures.
33 * Recovery procedures
50 * Auto recovery attempt is being done. Depends on:
52 - Auto-recovery configuration
80 * Configure reporter's generic parameters (like: disable/enable auto recovery)
81 * Invoke recovery procedure
95 - Triggers reporter's recovery procedure.
98 event in terms of recovery flow should follow closely that of a real
/linux/include/linux/
H A Ddm-region-hash.h87 * Region recovery control.
90 /* Prepare some regions for recovery by starting to quiesce them. */
93 /* Try fetching a quiesced region for recovery. */
96 /* Report recovery end on a region. */
99 /* Returns number of regions with recovery work outstanding. */
102 /* Start/stop recovery. */
/linux/fs/ocfs2/
H A Djournal.c81 * offline slots during recovery and mount
206 /* disable any new recovery threads and wait for any currently in ocfs2_recovery_exit()
213 /* At this point, we know that no more recovery threads can be in ocfs2_recovery_exit()
214 * launched, so wait for any recovery completion work to in ocfs2_recovery_exit()
220 * Now that recovery is shut down, and the osb is about to be in ocfs2_recovery_exit()
924 /* Skip recovery waits here - journal inode metadata never in ocfs2_journal_init()
1289 /* Does the second half of the recovery process. By this point, the
1291 * hence it's no longer in the recovery map, but there's still some
1292 * cleanup we can do which shouldn't happen within the recovery thread
1296 * NOTE: This function can and will sleep on recovery of other nodes
[all …]
/linux/Documentation/i2c/
H A Dgpio-fault-injection.rst36 master under test should detect this condition and trigger a bus recovery (see
38 core (see 'struct bus_recovery_info'). However, the bus recovery will not
47 device. Bus recovery should be able to fix these situations. But please note:
51 and will init a bus recovery on its own. If you want to implement bus recovery
64 recovery. This time, however, it should succeed and the device should release
82 This is why bus recovery (up to 9 clock pulses) must either check SDA or send
/linux/net/ipv4/
H A Dtcp_recovery.c11 * the recovery or starting the recovery by DUPACK threshold. in tcp_rack_reo_wnd()
131 * the aRTT term is bounded by the fast recovery or timeout, in tcp_rack_advance()
172 * triggering fast recovery, increment reo_wnd by min_rtt/4 (upper bounded
178 * recovery undo). After that, reset it to default (min_rtt/4).
212 /* RFC6582 NewReno recovery for non-SACK connection. It simply retransmits
214 * a) three or more DUPACKs to start the fast recovery
215 * b) an ACK acknowledging new data during the fast recovery.
/linux/include/linux/qed/
H A Dqede_rdma.h53 int qede_rdma_dev_add(struct qede_dev *dev, bool recovery);
56 void qede_rdma_dev_remove(struct qede_dev *dev, bool recovery);
62 bool recovery) in qede_rdma_dev_add() argument
70 bool recovery) {} in qede_rdma_dev_remove() argument
/linux/drivers/md/
H A Dmd.c518 set_bit(MD_RECOVERY_NEEDED, &mddev->recovery); in __mddev_resume()
1413 /* active but not in sync implies recovery up to in super_90_validate()
1997 &mddev->recovery)) in super_1_validate()
2460 /* May as well allow recovery to be retried once */ in bind_rdev_to_array()
2656 * curr_resync_completed can only be used during recovery. in md_update_sb()
2663 test_bit(MD_RECOVERY_RUNNING, &mddev->recovery) && in md_update_sb()
2664 test_bit(MD_RECOVERY_RECOVER, &mddev->recovery) && in md_update_sb()
2665 !test_bit(MD_RECOVERY_RESHAPE, &mddev->recovery) && in md_update_sb()
2713 * then a recovery will happen and soon that array won't in md_update_sb()
2794 if (test_bit(MD_RECOVERY_RUNNING, &mddev->recovery)) in md_update_sb()
[all …]
H A Draid1.h32 * 2/ when resync/recovery is known to be happening - i.e. in code that is
33 * called as part of performing resync/recovery.
79 * whether to retry a read, writeout a resync or recovery
93 * We don't allow both normal IO and resync/recovery IO at
94 * the same time - resync/recovery can only happen when there
113 * recovery to be attempted as we expect a read error.
/linux/Documentation/devicetree/bindings/sound/
H A Dst,sta32x.txt44 If present, thermal warning recovery is enabled.
46 - st,fault-detect-recovery:
47 If present, fault detect recovery is enabled.
52 - st,fault-detect-recovery:
53 If present, then fault recovery will be enabled.
/linux/arch/powerpc/platforms/powernv/
H A Dopal-hmi.c47 "Core checkstop during recovery" }, in print_core_checkstop_reason()
54 "Recovery in maintenance mode" }, in print_core_checkstop_reason()
64 "Hang Recovery Failed (core check stop)" }, in print_core_checkstop_reason()
203 "Processor Recovery done", in print_hmi_event_info()
204 "Processor recovery occurred again", in print_hmi_event_info()
205 "Processor recovery occurred for masked error", in print_hmi_event_info()
211 "SCOM has set a reserved FIR bit to cause recovery", in print_hmi_event_info()
212 "Debug trigger has set a reserved FIR bit to cause recovery", in print_hmi_event_info()
214 "CAPP recovery process is in progress", in print_hmi_event_info()
/linux/arch/s390/pci/
H A Dpci_event.c94 pr_info("%s: Automatic recovery failed after initial reporting\n", pci_name(pdev)); in zpci_event_notify_error_detected()
119 pr_info("%s: Automatic recovery failed after MMIO re-enable\n", in zpci_event_do_error_state_clear()
154 pr_info("%s: Automatic recovery failed after slot reset\n", pci_name(pdev)); in zpci_event_do_reset()
164 * We follow the scheme outlined in Documentation/PCI/pci-error-recovery.rst.
165 * With the simplification that recovery always happens per function
177 * configuration space while we perform recovery. in zpci_event_attempt_error_recovery()
198 pr_info("%s: The %s driver bound to the device does not support error recovery\n", in zpci_event_attempt_error_recovery()
218 pr_err("%s: Automatic recovery failed; operator intervention is required\n", in zpci_event_attempt_error_recovery()
287 case 0x0040: /* Service Action or Error Recovery Failed */ in __zpci_event_error()
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Dsetup.c193 /* ICOSQ recovery deactivates RQs. Suspend the recovery to avoid in mlx5e_activate_xsk()
194 * activating XSKRQ in the middle of recovery. in mlx5e_activate_xsk()
205 /* ICOSQ recovery may reactivate XSKRQ if clear_bit is called in the in mlx5e_deactivate_xsk()
206 * middle of recovery. Suspend the recovery to avoid it. in mlx5e_deactivate_xsk()
/linux/drivers/net/ethernet/qlogic/qede/
H A Dqede_rdma.c28 /* Leftovers from previous error recovery */ in _qede_rdma_dev_add()
86 int qede_rdma_dev_add(struct qede_dev *edev, bool recovery) in qede_rdma_dev_add() argument
94 if (recovery) in qede_rdma_dev_add()
116 void qede_rdma_dev_remove(struct qede_dev *edev, bool recovery) in qede_rdma_dev_remove() argument
122 if (!recovery) { in qede_rdma_dev_remove()
219 /* If device has experienced recovery it was already removed */ in qede_rdma_unregister_driver()
310 /* If a recovery was experienced avoid adding the event */ in qede_rdma_add_event()
/linux/tools/testing/selftests/powerpc/eeh/
H A Deeh-functions.sh72 # The ahci driver doesn't support error recovery. If the ahci device
77 log "$dev, Skipped: ahci doesn't support recovery"
83 # and injecting new errors during the recovery process will probably
84 # result in the recovery failing and the device being marked as
117 # The current record holder for longest recovery time is:
145 # error recovery
/linux/fs/xfs/
H A Dxfs_buf_item_recover.c31 * recovery.
39 * This structure is used during recovery to record the buf log items which
140 /* log buffer item recovery */
143 * Sort buffer items for log recovery. Most buffer items should end up on the
227 * We can only do post recovery validation on items on CRC enabled in xlog_recover_validate_buf_type()
423 * the buffer. This can occur if log recovery is interrupted at some in xlog_recover_validate_buf_type()
425 * subsequent mount starts recovery from the beginning. in xlog_recover_validate_buf_type()
442 * Perform a 'normal' buffer recovery. Each logged region of the
533 * Perform a dquot buffer recovery.
536 * Else, treat it as a regular buffer and do recovery.
[all …]

12345678910>>...55