| /* |
| * Copyright (c) 2013, NVIDIA CORPORATION. All rights reserved. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program. If not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #include <linux/sysedp.h> |
| #include <linux/platform_device.h> |
| #include <linux/workqueue.h> |
| #include <linux/kernel.h> |
| #include <linux/interrupt.h> |
| |
| #define STATE_MAX_MW 20000 |
| #define STATE_STEP_MW 500 |
| #define NSTATES (STATE_MAX_MW / STATE_STEP_MW + 1) |
| |
| inline unsigned int count_state(int mw) |
| { |
| int state; |
| state = mw > 0 ? mw / STATE_STEP_MW + 1 : 0; |
| return min(state, NSTATES - 1); |
| } |
| |
| static void oc_throttle_alarm(struct sysedp_reactive_capping_platform_data *h) |
| { |
| mutex_lock(&h->mutex); |
| |
| h->cur_capping_mw += h->step_alarm_mw; |
| h->cur_capping_mw = min(h->cur_capping_mw, h->max_capping_mw); |
| |
| cancel_delayed_work(&h->work); |
| |
| sysedp_set_state(&h->sysedpc, count_state(h->cur_capping_mw)); |
| |
| schedule_delayed_work(&h->work, msecs_to_jiffies(h->relax_ms)); |
| |
| mutex_unlock(&h->mutex); |
| } |
| |
| static void oc_throttle_work(struct work_struct *work) |
| { |
| struct sysedp_reactive_capping_platform_data *h; |
| h = container_of(to_delayed_work(work), |
| struct sysedp_reactive_capping_platform_data, |
| work); |
| mutex_lock(&h->mutex); |
| h->cur_capping_mw -= h->step_relax_mw; |
| h->cur_capping_mw = max(h->cur_capping_mw, 0); |
| |
| sysedp_set_state(&h->sysedpc, count_state(h->cur_capping_mw)); |
| |
| if (h->cur_capping_mw) |
| schedule_delayed_work(&h->work, msecs_to_jiffies(h->relax_ms)); |
| |
| mutex_unlock(&h->mutex); |
| } |
| |
| static irqreturn_t sysedp_reactive_capping_irq_handler(int irq, void *data) |
| { |
| if (!data) |
| return IRQ_NONE; |
| |
| oc_throttle_alarm(data); |
| return IRQ_HANDLED; |
| } |
| |
| |
| static unsigned int capping_states[NSTATES]; |
| |
| static int sysedp_reactive_capping_probe(struct platform_device *pdev) |
| { |
| int ret, i; |
| struct sysedp_reactive_capping_platform_data *pdata; |
| |
| pdata = pdev->dev.platform_data; |
| |
| if (!pdata) |
| return -EINVAL; |
| |
| /* update static capping_states table */ |
| for (i = 0; i < NSTATES; i++) |
| capping_states[i] = i * STATE_STEP_MW; |
| |
| /* sysedpc consumer name must be initialized */ |
| if (pdata->sysedpc.name[0] == '\0') |
| return -EINVAL; |
| pdata->sysedpc.states = capping_states; |
| pdata->sysedpc.num_states = ARRAY_SIZE(capping_states); |
| ret = sysedp_register_consumer(&pdata->sysedpc); |
| if (ret) { |
| pr_err("sysedp_reactive_capping_probe: consumer register failed (%d)\n", |
| ret); |
| return ret; |
| } |
| mutex_init(&pdata->mutex); |
| INIT_DELAYED_WORK(&pdata->work, oc_throttle_work); |
| |
| ret = request_threaded_irq(pdata->irq, |
| NULL, |
| sysedp_reactive_capping_irq_handler, |
| pdata->irq_flags, |
| pdata->sysedpc.name, |
| pdata); |
| if (ret) { |
| pr_err("sysedp_reactive_capping_probe: request_threaded_irq failed (%d)\n", |
| ret); |
| sysedp_unregister_consumer(&pdata->sysedpc); |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| static struct platform_driver sysedp_reactive_capping_driver = { |
| .probe = sysedp_reactive_capping_probe, |
| .driver = { |
| .name = "sysedp_reactive_capping", |
| .owner = THIS_MODULE |
| } |
| }; |
| |
| static __init int sysedp_reactive_capping_init(void) |
| { |
| return platform_driver_register(&sysedp_reactive_capping_driver); |
| } |
| late_initcall(sysedp_reactive_capping_init); |