/linux-6.14.4/arch/arm64/kvm/ |
D | vmid.c | 63 * Unlike ASID allocator, we expect less frequent rollover in in flush_context() 67 * invalidation over the inner shareable domain on rollover. in flush_context() 151 * reserving the VMID space needlessly on rollover. in kvm_arm_vmid_update() 153 * handle the sync with a concurrent rollover. in kvm_arm_vmid_update() 180 * Expect allocation after rollover to fail if we don't have in kvm_arm_vmid_alloc_init()
|
/linux-6.14.4/arch/arm64/mm/ |
D | context.c | 116 * rollover, but hasn't run another task in in flush_context() 168 * If our current ASID was active during a rollover, we in new_context() 230 * cmpxchg. Racing with a concurrent rollover means that either: in check_and_switch_context() 233 * lock. Taking the lock synchronises with the rollover and so in check_and_switch_context() 295 * We went through one or more rollover since that ASID was in arm64_mm_context_get() 384 * Expect allocation after rollover to fail if we don't have at least in asids_update_limit() 392 * There must always be an ASID available after rollover. Ensure that, in asids_update_limit()
|
/linux-6.14.4/drivers/input/misc/ |
D | rotary_encoder.c | 41 bool rollover; member 84 if (encoder->rollover) in rotary_encoder_report_event() 90 if (encoder->rollover || pos < encoder->steps) in rotary_encoder_report_event() 94 if (encoder->rollover) in rotary_encoder_report_event() 210 encoder->rollover = in rotary_encoder_probe() 211 device_property_read_bool(dev, "rotary-encoder,rollover"); in rotary_encoder_probe()
|
/linux-6.14.4/arch/mips/mm/ |
D | context.c | 80 * rollover, but hasn't run another task in in flush_context() 137 * If our current MMID was active during a rollover, we in get_new_mmid() 197 * asid_cache with a relaxed cmpxchg. Racing with a concurrent rollover in check_switch_mmu_context() 202 * with the rollover and so we are forced to see the updated in check_switch_mmu_context() 273 * Expect allocation after rollover to fail if we don't have at least in mmid_init()
|
/linux-6.14.4/arch/arc/include/asm/ |
D | mmu_context.h | 36 * A new allocation cycle, post rollover, could potentially reassign an ASID 40 * automagically increments the generation when lower 8 bits rollover. 79 /* move to new ASID and handle rollover */ in get_new_mmu_context() 85 * Above check for rollover of 8 bit ASID in 32 bit container. in get_new_mmu_context()
|
/linux-6.14.4/tools/testing/selftests/net/tcp_ao/ |
D | seq-ext.c | 127 synchronize_threads(); /* 5: verify the connection during SEQ-number rollover */ in server_fn() 138 synchronize_threads(); /* 6: verify counters after SEQ-number rollover */ in server_fn() 218 synchronize_threads(); /* 5: verify the connection during SEQ-number rollover */ in client_fn() 224 synchronize_threads(); /* 5: verify counters after SEQ-number rollover */ in client_fn()
|
/linux-6.14.4/Documentation/devicetree/bindings/input/ |
D | rotary-encoder.yaml | 43 rotary-encoder,rollover: 46 Automatic rollover when the rotary value becomes
|
/linux-6.14.4/arch/csky/include/asm/ |
D | asid.h | 54 * cmpxchg. Racing with a concurrent rollover means that either: in asid_check_context() 57 * lock. Taking the lock synchronises with the rollover and so in asid_check_context()
|
/linux-6.14.4/arch/csky/mm/ |
D | asid.c | 36 * rollover, but hasn't run another task in in flush_context() 90 * If our current ASID was active during a rollover, we in new_context() 176 * Expect allocation after rollover to fail if we don't have at least in asid_allocator_init()
|
/linux-6.14.4/arch/riscv/mm/ |
D | context.c | 75 * If this CPU has already been through a rollover, but in __flush_context() 107 * If our current CONTEXT was active during a rollover, we in __new_context() 157 * Following is how we handle racing with a concurrent rollover: in set_mm_asid() 160 * lock. Taking the lock synchronises with the rollover and so in set_mm_asid()
|
/linux-6.14.4/arch/arm64/include/asm/ |
D | mmu.h | 34 * rollover event (see new_context() and flush_context()). In this case, 46 * | <rollover>
|
/linux-6.14.4/drivers/rtc/ |
D | interface.c | 302 /* For simplicity, only support date rollover for now */ in __rtc_read_alarm() 325 /* with luck, no rollover is needed */ in __rtc_read_alarm() 332 /* 24 hour rollover ... if it's now 10am Monday, an alarm that in __rtc_read_alarm() 338 dev_dbg(&rtc->dev, "alarm rollover: %s\n", "day"); in __rtc_read_alarm() 343 /* Month rollover ... if it's the 31th, an alarm on the 3rd will in __rtc_read_alarm() 349 dev_dbg(&rtc->dev, "alarm rollover: %s\n", "month"); in __rtc_read_alarm() 362 /* Year rollover ... easy except for leap years! */ in __rtc_read_alarm() 364 dev_dbg(&rtc->dev, "alarm rollover: %s\n", "year"); in __rtc_read_alarm() 372 dev_warn(&rtc->dev, "alarm rollover not handled\n"); in __rtc_read_alarm()
|
/linux-6.14.4/arch/arm/mm/ |
D | context.c | 81 * any issues across a rollover). 147 * rollover, but hasn't run another task in in flush_context() 199 * If our current ASID was active during a rollover, we in new_context()
|
/linux-6.14.4/drivers/net/ethernet/ti/ |
D | cpts.h | 82 CPTS_EV_ROLL, /* Time Stamp Rollover Event */ 83 CPTS_EV_HALF, /* Time Stamp Half Rollover Event */
|
/linux-6.14.4/drivers/net/ethernet/marvell/octeontx2/af/ |
D | ptp.c | 162 /* check nsec rollover */ in read_ptp_tstmp_sec_nsec() 184 * Issue #1: At the time of 1 sec rollover of the nano-second counter, in ptp_calc_adjusted_comp() 220 /* slow down the ptp clock to not rollover early */ in ptp_calc_adjusted_comp() 387 /* Program the seconds rollover value to 1 second */ in ptp_start()
|
/linux-6.14.4/net/packet/ |
D | af_packet.c | 1392 u32 *history = po->rollover->history; in fanout_flow_is_huge() 1458 i = j = min_t(int, po->rollover->sock, num - 1); in fanout_demux_rollover() 1465 po->rollover->sock = i; in fanout_demux_rollover() 1466 atomic_long_inc(&po->rollover->num); in fanout_demux_rollover() 1468 atomic_long_inc(&po->rollover->num_huge); in fanout_demux_rollover() 1476 atomic_long_inc(&po->rollover->num_failed); in fanout_demux_rollover() 1732 struct packet_rollover *rollover = NULL; in fanout_add() local 1767 rollover = kzalloc(sizeof(*rollover), GFP_KERNEL); in fanout_add() 1768 if (!rollover) in fanout_add() 1770 atomic_long_set(&rollover->num, 0); in fanout_add() [all …]
|
D | internal.h | 123 struct packet_rollover *rollover; member
|
/linux-6.14.4/Documentation/hwmon/ |
D | fam15h_power.rst | 113 if (Jy < Jx) // Rollover has occurred
|
/linux-6.14.4/drivers/mtd/devices/ |
D | mchp48l640.c | 227 * we set PRO bit (page rollover), but writing length > page size in mchp48l640_write() 285 * we set PRO bit (page rollover), but if read length > page size in mchp48l640_read()
|
/linux-6.14.4/tools/testing/selftests/timers/ |
D | set-2038.c | 117 /* Test rollover behavior 32bit edge */ in main()
|
/linux-6.14.4/samples/pfsm/ |
D | pfsm-wakeup.c | 62 /* Set RTC alarm to ALARM_DELTA_SEC sec in the future, and check for rollover */ in main()
|
/linux-6.14.4/sound/isa/gus/ |
D | gus_pcm.c | 111 /* enable RAMP IRQ + rollover */ in snd_gf1_pcm_trigger_up() 115 ramp_ctrl &= ~0x04; /* disable rollover */ in snd_gf1_pcm_trigger_up() 203 ramp_ctrl |= 0x04; /* enable rollover */ in snd_gf1_pcm_interrupt_wave() 254 /* stop ramp, but leave rollover bit untouched */ in snd_gf1_pcm_interrupt_volume()
|
/linux-6.14.4/Documentation/ABI/testing/ |
D | sysfs-bus-pci-devices-aer_stats | 26 RELAY_NUM Rollover 0
|
/linux-6.14.4/drivers/usb/dwc2/ |
D | hcd.h | 603 * is done modulo FRLISTEN_64_SIZE. This accounts for the rollover of the 616 * done modulo HFNUM_MAX_FRNUM. This accounts for the rollover of the 626 * modulo HFNUM_MAX_FRNUM. This accounts for the rollover of the frame
|
/linux-6.14.4/drivers/gpu/drm/amd/include/ |
D | kgd_pp_interface.h | 804 /* PCIE replay rollover accumulated count */ 877 /* PCIE replay rollover accumulated count */ 961 /* PCIE replay rollover accumulated count */ 1055 /* PCIE replay rollover accumulated count */
|