blob: c2f118f6b77690972f2b451348e1093ea08f9ad1 [file] [log] [blame]
// SPDX-License-Identifier: GPL-2.0
/*
* Abrolhos Edge TPU ML accelerator device host support.
*
* Copyright (C) 2019 Google, Inc.
*/
#include <linux/irqreturn.h>
#include "abrolhos-platform.h"
#include "edgetpu-config.h"
#include "edgetpu-debug-dump.h"
#include "edgetpu-internal.h"
#include "edgetpu-mailbox.h"
#include "edgetpu-telemetry.h"
#include "edgetpu-wakelock.h"
#include "mobile-pm.h"
#define HOST_NONSECURE_INTRSRCMASKREG 0x000f0004
#define SSMT_NS_READ_STREAM_VID_OFFSET(n) (0x1000u + (0x4u * (n)))
#define SSMT_NS_WRITE_STREAM_VID_OFFSET(n) (0x1200u + (0x4u * (n)))
#define SSMT_NS_READ_STREAM_VID_REG(base, n) \
((base) + SSMT_NS_READ_STREAM_VID_OFFSET(n))
#define SSMT_NS_WRITE_STREAM_VID_REG(base, n) \
((base) + SSMT_NS_WRITE_STREAM_VID_OFFSET(n))
/*
* The interrupt handler for mailboxes.
*
* This handler reads the IntSrcStatusReg available on Abrolhos to get the
* response queue doorbell status of mailboxes 0-7 in one read. It loops
* through the bits to check pending interrupts and invokes their IRQ
* handlers. This allows for more efficient handling of interrupts than
* edgetpu_mailbox_handle_irq handler.
*/
static irqreturn_t
abrolhos_mailbox_handle_irq(struct edgetpu_mailbox_manager *mgr)
{
struct edgetpu_mailbox *mailbox;
uint i;
u32 val, clear_val;
u8 mailbox_rsp_queue_status_bits;
if (!mgr)
return IRQ_NONE;
read_lock(&mgr->mailboxes_lock);
val = edgetpu_dev_read_32(mgr->etdev,
HOST_NONSECURE_INT_SRC_STATUS_REG);
clear_val = val & (0xff << 1);
edgetpu_dev_write_32(mgr->etdev,
HOST_NONSECURE_INT_SRC_CLEAR_REG, clear_val);
mailbox_rsp_queue_status_bits = val >> 1;
for (i = 0; i < mgr->num_mailbox; i++) {
mailbox = mgr->mailboxes[i];
if (mailbox && (mailbox_rsp_queue_status_bits & 0x1) &&
mailbox->handle_irq)
mailbox->handle_irq(mailbox);
mailbox_rsp_queue_status_bits >>= 1;
if (!mailbox_rsp_queue_status_bits)
break;
}
read_unlock(&mgr->mailboxes_lock);
return IRQ_HANDLED;
}
irqreturn_t edgetpu_chip_irq_handler(int irq, void *arg)
{
struct edgetpu_dev *etdev = arg;
edgetpu_telemetry_irq_handler(etdev);
edgetpu_debug_dump_resp_handler(etdev);
return abrolhos_mailbox_handle_irq(etdev->mailbox_manager);
}
u64 edgetpu_chip_tpu_timestamp(struct edgetpu_dev *etdev)
{
return edgetpu_dev_read_64(etdev, EDGETPU_REG_CPUNS_TIMESTAMP);
}
void edgetpu_chip_init(struct edgetpu_dev *etdev)
{
int i;
struct edgetpu_mobile_platform_dev *etmdev = to_mobile_dev(etdev);
/* Disable the CustomBlock Interrupt. */
edgetpu_dev_write_32(etdev, HOST_NONSECURE_INTRSRCMASKREG, 0x1);
if (!etmdev->ssmt_base)
return;
/* Setup non-secure SCIDs, assume VID = SCID */
for (i = 0; i < EDGETPU_NCONTEXTS; i++) {
writel(i, SSMT_NS_READ_STREAM_VID_REG(etmdev->ssmt_base, i));
writel(i, SSMT_NS_WRITE_STREAM_VID_REG(etmdev->ssmt_base, i));
}
}
void edgetpu_chip_exit(struct edgetpu_dev *etdev)
{
}
void edgetpu_mark_probe_fail(struct edgetpu_dev *etdev)
{
}
void edgetpu_chip_handle_reverse_kci(struct edgetpu_dev *etdev,
struct edgetpu_kci_response_element *resp)
{
switch (resp->code) {
case RKCI_CODE_PM_QOS:
mobile_pm_set_pm_qos(etdev, resp->retval);
break;
case RKCI_CODE_BTS:
mobile_pm_set_bts(etdev, resp->retval);
break;
default:
etdev_warn(etdev, "%s: Unrecognized KCI request: %u\n",
__func__, resp->code);
break;
}
}
static int abrolhos_check_ext_mailbox_args(const char *func,
struct edgetpu_dev *etdev,
struct edgetpu_ext_mailbox_ioctl *args)
{
if (args->type != EDGETPU_EXT_MAILBOX_TYPE_TZ) {
etdev_err(etdev, "%s: Invalid type %d != %d\n", func,
args->type, EDGETPU_EXT_MAILBOX_TYPE_TZ);
return -EINVAL;
}
if (args->count != 1) {
etdev_err(etdev, "%s: Invalid mailbox count: %d != 1\n", func,
args->count);
return -EINVAL;
}
return 0;
}
int edgetpu_chip_acquire_ext_mailbox(struct edgetpu_client *client,
struct edgetpu_ext_mailbox_ioctl *args)
{
struct edgetpu_mobile_platform_dev *etmdev = to_mobile_dev(client->etdev);
int ret;
ret = abrolhos_check_ext_mailbox_args(__func__, client->etdev,
args);
if (ret)
return ret;
mutex_lock(&etmdev->tz_mailbox_lock);
if (etmdev->secure_client) {
etdev_err(client->etdev,
"TZ mailbox already in use by PID %d\n",
etmdev->secure_client->pid);
mutex_unlock(&etmdev->tz_mailbox_lock);
return -EBUSY;
}
ret = edgetpu_mailbox_enable_ext(client, ABROLHOS_TZ_MAILBOX_ID, NULL);
if (!ret)
etmdev->secure_client = client;
mutex_unlock(&etmdev->tz_mailbox_lock);
return ret;
}
int edgetpu_chip_release_ext_mailbox(struct edgetpu_client *client,
struct edgetpu_ext_mailbox_ioctl *args)
{
struct edgetpu_mobile_platform_dev *etmdev = to_mobile_dev(client->etdev);
int ret = 0;
ret = abrolhos_check_ext_mailbox_args(__func__, client->etdev,
args);
if (ret)
return ret;
mutex_lock(&etmdev->tz_mailbox_lock);
if (!etmdev->secure_client) {
etdev_warn(client->etdev, "TZ mailbox already released\n");
mutex_unlock(&etmdev->tz_mailbox_lock);
return 0;
}
if (etmdev->secure_client != client) {
etdev_err(client->etdev,
"TZ mailbox owned by different client\n");
mutex_unlock(&etmdev->tz_mailbox_lock);
return -EBUSY;
}
etmdev->secure_client = NULL;
ret = edgetpu_mailbox_disable_ext(client, ABROLHOS_TZ_MAILBOX_ID);
mutex_unlock(&etmdev->tz_mailbox_lock);
return ret;
}
void edgetpu_chip_client_remove(struct edgetpu_client *client)
{
struct edgetpu_mobile_platform_dev *etmdev = to_mobile_dev(client->etdev);
mutex_lock(&etmdev->tz_mailbox_lock);
if (etmdev->secure_client == client) {
etmdev->secure_client = NULL;
edgetpu_mailbox_disable_ext(client, ABROLHOS_TZ_MAILBOX_ID);
}
mutex_unlock(&etmdev->tz_mailbox_lock);
}