Lines Matching +full:led +full:- +full:cur
1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Copyright (C) 2006 - 2007 Ivo van Doorn
96 * rfkill method calls -- which will cause an AB-BA deadlock situation.
98 * To fix that, we need to rework this code here to be mostly lock-free
113 bool cur, sav; member
124 if (!rfkill->registered) in rfkill_led_trigger_event()
127 trigger = &rfkill->led_trigger; in rfkill_led_trigger_event()
129 if (rfkill->state & RFKILL_BLOCK_ANY) in rfkill_led_trigger_event()
135 static int rfkill_led_trigger_activate(struct led_classdev *led) in rfkill_led_trigger_activate() argument
139 rfkill = container_of(led->trigger, struct rfkill, led_trigger); in rfkill_led_trigger_activate()
148 return rfkill->led_trigger.name; in rfkill_get_led_trigger_name()
156 rfkill->ledtrigname = name; in rfkill_set_led_trigger_name()
162 rfkill->led_trigger.name = rfkill->ledtrigname in rfkill_led_trigger_register()
163 ? : dev_name(&rfkill->dev); in rfkill_led_trigger_register()
164 rfkill->led_trigger.activate = rfkill_led_trigger_activate; in rfkill_led_trigger_register()
165 return led_trigger_register(&rfkill->led_trigger); in rfkill_led_trigger_register()
170 led_trigger_unregister(&rfkill->led_trigger); in rfkill_led_trigger_unregister()
184 if (!(rfkill->state & RFKILL_BLOCK_ANY)) { in rfkill_global_led_trigger_worker()
208 rfkill_any_led_trigger.name = "rfkill-any"; in rfkill_global_led_trigger_register()
213 rfkill_none_led_trigger.name = "rfkill-none"; in rfkill_global_led_trigger_register()
264 ev->idx = rfkill->idx; in rfkill_fill_event()
265 ev->type = rfkill->type; in rfkill_fill_event()
266 ev->op = op; in rfkill_fill_event()
268 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_fill_event()
269 ev->hard = !!(rfkill->state & RFKILL_BLOCK_HW); in rfkill_fill_event()
270 ev->soft = !!(rfkill->state & (RFKILL_BLOCK_SW | in rfkill_fill_event()
272 ev->hard_block_reasons = rfkill->hard_block_reasons; in rfkill_fill_event()
273 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_fill_event()
285 rfkill_fill_event(&ev->ev, rfkill, op); in rfkill_send_events()
286 mutex_lock(&data->mtx); in rfkill_send_events()
287 list_add_tail(&ev->list, &data->events); in rfkill_send_events()
288 mutex_unlock(&data->mtx); in rfkill_send_events()
289 wake_up_interruptible(&data->read_wait); in rfkill_send_events()
295 if (!rfkill->registered) in rfkill_event()
298 kobject_uevent(&rfkill->dev.kobj, KOBJ_CHANGE); in rfkill_event()
305 * rfkill_set_block - wrapper for set_block method
319 if (unlikely(rfkill->dev.power.power_state.event & PM_EVENT_SLEEP)) in rfkill_set_block()
324 * _hard_ kill state -- whenever something tries to change the in rfkill_set_block()
327 if (rfkill->ops->query) in rfkill_set_block()
328 rfkill->ops->query(rfkill, rfkill->data); in rfkill_set_block()
330 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_set_block()
331 prev = rfkill->state & RFKILL_BLOCK_SW; in rfkill_set_block()
334 rfkill->state |= RFKILL_BLOCK_SW_PREV; in rfkill_set_block()
336 rfkill->state &= ~RFKILL_BLOCK_SW_PREV; in rfkill_set_block()
339 rfkill->state |= RFKILL_BLOCK_SW; in rfkill_set_block()
341 rfkill->state &= ~RFKILL_BLOCK_SW; in rfkill_set_block()
343 rfkill->state |= RFKILL_BLOCK_SW_SETCALL; in rfkill_set_block()
344 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_set_block()
346 err = rfkill->ops->set_block(rfkill->data, blocked); in rfkill_set_block()
348 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_set_block()
351 * Failed -- reset status to _PREV, which may be different in rfkill_set_block()
355 if (rfkill->state & RFKILL_BLOCK_SW_PREV) in rfkill_set_block()
356 rfkill->state |= RFKILL_BLOCK_SW; in rfkill_set_block()
358 rfkill->state &= ~RFKILL_BLOCK_SW; in rfkill_set_block()
360 rfkill->state &= ~RFKILL_BLOCK_SW_SETCALL; in rfkill_set_block()
361 rfkill->state &= ~RFKILL_BLOCK_SW_PREV; in rfkill_set_block()
362 curr = rfkill->state & RFKILL_BLOCK_SW; in rfkill_set_block()
363 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_set_block()
376 if (!rfkill->need_sync) in rfkill_sync()
379 rfkill_set_block(rfkill, rfkill_global_states[rfkill->type].cur); in rfkill_sync()
380 rfkill->need_sync = false; in rfkill_sync()
388 rfkill_global_states[type].cur = blocked; in rfkill_update_global_state()
393 rfkill_global_states[i].cur = blocked; in rfkill_update_global_state()
400 * __rfkill_switch_all - Toggle state of all switches of given type
415 if (rfkill->type != type && type != RFKILL_TYPE_ALL) in __rfkill_switch_all()
423 * rfkill_switch_all - Toggle state of all switches of given type
446 * rfkill_epo - emergency power off all transmitters
448 * This kicks all non-suspended rfkill devices to RFKILL_STATE_SOFT_BLOCKED,
449 * ignoring everything in its path but rfkill_global_mutex and rfkill->mutex.
469 rfkill_global_states[i].sav = rfkill_global_states[i].cur; in rfkill_epo()
470 rfkill_global_states[i].cur = true; in rfkill_epo()
477 * rfkill_restore_states - restore global states
499 * rfkill_remove_epo_lock - unlock state changes
501 * Used by rfkill-input manually unlock state changes, when
515 * rfkill_is_epo_lock_active - returns true EPO is active
529 * rfkill_get_global_sw_state - returns global state for a type
537 return rfkill_global_states[type].cur; in rfkill_get_global_sw_state()
550 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_set_hw_state_reason()
551 prev = !!(rfkill->hard_block_reasons & reason); in rfkill_set_hw_state_reason()
553 rfkill->state |= RFKILL_BLOCK_HW; in rfkill_set_hw_state_reason()
554 rfkill->hard_block_reasons |= reason; in rfkill_set_hw_state_reason()
556 rfkill->hard_block_reasons &= ~reason; in rfkill_set_hw_state_reason()
557 if (!rfkill->hard_block_reasons) in rfkill_set_hw_state_reason()
558 rfkill->state &= ~RFKILL_BLOCK_HW; in rfkill_set_hw_state_reason()
560 ret = !!(rfkill->state & RFKILL_BLOCK_ANY); in rfkill_set_hw_state_reason()
561 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_set_hw_state_reason()
566 if (rfkill->registered && prev != blocked) in rfkill_set_hw_state_reason()
567 schedule_work(&rfkill->uevent_work); in rfkill_set_hw_state_reason()
577 /* if in a ops->set_block right now, use other bit */ in __rfkill_set_sw_state()
578 if (rfkill->state & RFKILL_BLOCK_SW_SETCALL) in __rfkill_set_sw_state()
582 rfkill->state |= bit; in __rfkill_set_sw_state()
584 rfkill->state &= ~bit; in __rfkill_set_sw_state()
594 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_set_sw_state()
595 prev = !!(rfkill->state & RFKILL_BLOCK_SW); in rfkill_set_sw_state()
597 hwblock = !!(rfkill->state & RFKILL_BLOCK_HW); in rfkill_set_sw_state()
599 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_set_sw_state()
601 if (!rfkill->registered) in rfkill_set_sw_state()
605 schedule_work(&rfkill->uevent_work); in rfkill_set_sw_state()
619 BUG_ON(rfkill->registered); in rfkill_init_sw_state()
621 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_init_sw_state()
623 rfkill->persistent = true; in rfkill_init_sw_state()
624 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_init_sw_state()
635 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_set_states()
641 swprev = !!(rfkill->state & RFKILL_BLOCK_SW); in rfkill_set_states()
642 hwprev = !!(rfkill->state & RFKILL_BLOCK_HW); in rfkill_set_states()
645 rfkill->state |= RFKILL_BLOCK_HW; in rfkill_set_states()
647 rfkill->state &= ~RFKILL_BLOCK_HW; in rfkill_set_states()
649 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_set_states()
651 if (!rfkill->registered) { in rfkill_set_states()
652 rfkill->persistent = true; in rfkill_set_states()
655 schedule_work(&rfkill->uevent_work); in rfkill_set_states()
696 return sysfs_emit(buf, "%s\n", rfkill->name); in name_show()
705 return sysfs_emit(buf, "%s\n", rfkill_types[rfkill->type]); in type_show()
714 return sysfs_emit(buf, "%d\n", rfkill->idx); in index_show()
723 return sysfs_emit(buf, "%d\n", rfkill->persistent); in persistent_show()
732 return sysfs_emit(buf, "%d\n", (rfkill->state & RFKILL_BLOCK_HW) ? 1 : 0); in hard_show()
745 return sysfs_emit(buf, "%d\n", (rfkill->state & RFKILL_BLOCK_SW) ? 1 : 0); in soft_show()
756 return -EPERM; in soft_store()
763 return -EINVAL; in soft_store()
780 return sysfs_emit(buf, "0x%lx\n", rfkill->hard_block_reasons); in hard_block_reasons_show()
803 return sysfs_emit(buf, "%d\n", user_state_from_blocked(rfkill->state)); in state_show()
814 return -EPERM; in state_store()
822 return -EINVAL; in state_store()
861 error = add_uevent_var(env, "RFKILL_NAME=%s", rfkill->name); in rfkill_dev_uevent()
865 rfkill_types[rfkill->type]); in rfkill_dev_uevent()
868 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_dev_uevent()
869 state = rfkill->state; in rfkill_dev_uevent()
870 reasons = rfkill->hard_block_reasons; in rfkill_dev_uevent()
871 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_dev_uevent()
883 if (!rfkill->ops->poll) in rfkill_pause_polling()
886 rfkill->polling_paused = true; in rfkill_pause_polling()
887 cancel_delayed_work_sync(&rfkill->poll_work); in rfkill_pause_polling()
895 if (!rfkill->ops->poll) in rfkill_resume_polling()
898 rfkill->polling_paused = false; in rfkill_resume_polling()
900 if (rfkill->suspended) in rfkill_resume_polling()
904 &rfkill->poll_work, 0); in rfkill_resume_polling()
913 rfkill->suspended = true; in rfkill_suspend()
914 cancel_delayed_work_sync(&rfkill->poll_work); in rfkill_suspend()
922 bool cur; in rfkill_resume() local
924 rfkill->suspended = false; in rfkill_resume()
926 if (!rfkill->registered) in rfkill_resume()
929 if (!rfkill->persistent) { in rfkill_resume()
930 cur = !!(rfkill->state & RFKILL_BLOCK_SW); in rfkill_resume()
931 rfkill_set_block(rfkill, cur); in rfkill_resume()
934 if (rfkill->ops->poll && !rfkill->polling_paused) in rfkill_resume()
936 &rfkill->poll_work, 0); in rfkill_resume()
960 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_blocked()
961 state = rfkill->state; in rfkill_blocked()
962 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_blocked()
973 spin_lock_irqsave(&rfkill->lock, flags); in rfkill_soft_blocked()
974 state = rfkill->state; in rfkill_soft_blocked()
975 spin_unlock_irqrestore(&rfkill->lock, flags); in rfkill_soft_blocked()
993 if (WARN_ON(!ops->set_block)) in rfkill_alloc()
1006 spin_lock_init(&rfkill->lock); in rfkill_alloc()
1007 INIT_LIST_HEAD(&rfkill->node); in rfkill_alloc()
1008 rfkill->type = type; in rfkill_alloc()
1009 strcpy(rfkill->name, name); in rfkill_alloc()
1010 rfkill->ops = ops; in rfkill_alloc()
1011 rfkill->data = ops_data; in rfkill_alloc()
1013 dev = &rfkill->dev; in rfkill_alloc()
1014 dev->class = &rfkill_class; in rfkill_alloc()
1015 dev->parent = parent; in rfkill_alloc()
1029 * Poll hardware state -- driver will use one of the in rfkill_poll()
1033 rfkill->ops->poll(rfkill, rfkill->data); in rfkill_poll()
1036 &rfkill->poll_work, in rfkill_poll()
1067 return -EINVAL; in rfkill_register()
1069 dev = &rfkill->dev; in rfkill_register()
1073 if (rfkill->registered) { in rfkill_register()
1074 error = -EALREADY; in rfkill_register()
1078 rfkill->idx = rfkill_no; in rfkill_register()
1082 list_add_tail(&rfkill->node, &rfkill_list); in rfkill_register()
1092 rfkill->registered = true; in rfkill_register()
1094 INIT_DELAYED_WORK(&rfkill->poll_work, rfkill_poll); in rfkill_register()
1095 INIT_WORK(&rfkill->uevent_work, rfkill_uevent_work); in rfkill_register()
1096 INIT_WORK(&rfkill->sync_work, rfkill_sync_work); in rfkill_register()
1098 if (rfkill->ops->poll) in rfkill_register()
1100 &rfkill->poll_work, in rfkill_register()
1103 if (!rfkill->persistent || rfkill_epo_lock_active) { in rfkill_register()
1104 rfkill->need_sync = true; in rfkill_register()
1105 schedule_work(&rfkill->sync_work); in rfkill_register()
1108 bool soft_blocked = !!(rfkill->state & RFKILL_BLOCK_SW); in rfkill_register()
1111 __rfkill_switch_all(rfkill->type, soft_blocked); in rfkill_register()
1122 device_del(&rfkill->dev); in rfkill_register()
1124 list_del_init(&rfkill->node); in rfkill_register()
1135 if (rfkill->ops->poll) in rfkill_unregister()
1136 cancel_delayed_work_sync(&rfkill->poll_work); in rfkill_unregister()
1138 cancel_work_sync(&rfkill->uevent_work); in rfkill_unregister()
1139 cancel_work_sync(&rfkill->sync_work); in rfkill_unregister()
1141 rfkill->registered = false; in rfkill_unregister()
1143 device_del(&rfkill->dev); in rfkill_unregister()
1147 list_del_init(&rfkill->node); in rfkill_unregister()
1158 put_device(&rfkill->dev); in rfkill_destroy()
1170 return -ENOMEM; in rfkill_fop_open()
1172 data->max_size = RFKILL_EVENT_SIZE_V1; in rfkill_fop_open()
1174 INIT_LIST_HEAD(&data->events); in rfkill_fop_open()
1175 mutex_init(&data->mtx); in rfkill_fop_open()
1176 init_waitqueue_head(&data->read_wait); in rfkill_fop_open()
1189 rfkill_fill_event(&ev->ev, rfkill, RFKILL_OP_ADD); in rfkill_fop_open()
1190 mutex_lock(&data->mtx); in rfkill_fop_open()
1191 list_add_tail(&ev->list, &data->events); in rfkill_fop_open()
1192 mutex_unlock(&data->mtx); in rfkill_fop_open()
1194 list_add(&data->list, &rfkill_fds); in rfkill_fop_open()
1197 file->private_data = data; in rfkill_fop_open()
1203 mutex_destroy(&data->mtx); in rfkill_fop_open()
1204 list_for_each_entry_safe(ev, tmp, &data->events, list) in rfkill_fop_open()
1207 return -ENOMEM; in rfkill_fop_open()
1212 struct rfkill_data *data = file->private_data; in rfkill_fop_poll()
1215 poll_wait(file, &data->read_wait, wait); in rfkill_fop_poll()
1217 mutex_lock(&data->mtx); in rfkill_fop_poll()
1218 if (!list_empty(&data->events)) in rfkill_fop_poll()
1220 mutex_unlock(&data->mtx); in rfkill_fop_poll()
1228 struct rfkill_data *data = file->private_data; in rfkill_fop_read()
1233 mutex_lock(&data->mtx); in rfkill_fop_read()
1235 while (list_empty(&data->events)) { in rfkill_fop_read()
1236 if (file->f_flags & O_NONBLOCK) { in rfkill_fop_read()
1237 ret = -EAGAIN; in rfkill_fop_read()
1240 mutex_unlock(&data->mtx); in rfkill_fop_read()
1241 /* since we re-check and it just compares pointers, in rfkill_fop_read()
1244 ret = wait_event_interruptible(data->read_wait, in rfkill_fop_read()
1245 !list_empty(&data->events)); in rfkill_fop_read()
1246 mutex_lock(&data->mtx); in rfkill_fop_read()
1252 ev = list_first_entry(&data->events, struct rfkill_int_event, in rfkill_fop_read()
1255 sz = min_t(unsigned long, sizeof(ev->ev), count); in rfkill_fop_read()
1256 sz = min_t(unsigned long, sz, data->max_size); in rfkill_fop_read()
1258 if (copy_to_user(buf, &ev->ev, sz)) in rfkill_fop_read()
1259 ret = -EFAULT; in rfkill_fop_read()
1261 list_del(&ev->list); in rfkill_fop_read()
1264 mutex_unlock(&data->mtx); in rfkill_fop_read()
1271 struct rfkill_data *data = file->private_data; in rfkill_fop_write()
1277 if (count < RFKILL_EVENT_SIZE_V1 - 1) in rfkill_fop_write()
1278 return -EINVAL; in rfkill_fop_write()
1286 count = min_t(size_t, count, data->max_size); in rfkill_fop_write()
1288 return -EFAULT; in rfkill_fop_write()
1291 return -EINVAL; in rfkill_fop_write()
1299 if (rfkill->type == ev.type || in rfkill_fop_write()
1306 if (rfkill->idx == ev.idx && in rfkill_fop_write()
1307 (rfkill->type == ev.type || in rfkill_fop_write()
1313 ret = -EINVAL; in rfkill_fop_write()
1324 struct rfkill_data *data = file->private_data; in rfkill_fop_release()
1328 list_del(&data->list); in rfkill_fop_release()
1331 mutex_destroy(&data->mtx); in rfkill_fop_release()
1332 list_for_each_entry_safe(ev, tmp, &data->events, list) in rfkill_fop_release()
1336 if (data->input_handler) in rfkill_fop_release()
1349 struct rfkill_data *data = file->private_data; in rfkill_fop_ioctl()
1350 int ret = -ENOTTY; in rfkill_fop_ioctl()
1354 return -ENOTTY; in rfkill_fop_ioctl()
1356 mutex_lock(&data->mtx); in rfkill_fop_ioctl()
1360 if (!data->input_handler) { in rfkill_fop_ioctl()
1363 data->input_handler = true; in rfkill_fop_ioctl()
1370 ret = -EFAULT; in rfkill_fop_ioctl()
1374 ret = -EINVAL; in rfkill_fop_ioctl()
1377 data->max_size = size; in rfkill_fop_ioctl()
1383 mutex_unlock(&data->mtx); in rfkill_fop_ioctl()