Files
2025-08-12 22:16:57 +02:00

873 lines
23 KiB
C
Executable File

// SPDX-License-Identifier: GPL-2.0-only
/*
* Copyright (c) 2022-2024 Qualcomm Innovation Center, Inc. All rights reserved.
*/
#include <linux/completion.h>
#include <linux/gunyah.h>
#include <linux/module.h>
#include <linux/mutex.h>
#include <linux/notifier.h>
#include <linux/of.h>
#include <linux/of_irq.h>
#include <linux/platform_device.h>
#include <linux/miscdevice.h>
#include <linux/auxiliary_bus.h>
#include <asm/gunyah.h>
#include "rsc_mgr.h"
#include "vm_mgr.h"
/* clang-format off */
#define RM_RPC_API_VERSION_MASK GENMASK(3, 0)
#define RM_RPC_HEADER_WORDS_MASK GENMASK(7, 4)
#define RM_RPC_API_VERSION FIELD_PREP(RM_RPC_API_VERSION_MASK, 1)
#define RM_RPC_HEADER_WORDS FIELD_PREP(RM_RPC_HEADER_WORDS_MASK, \
(sizeof(struct gunyah_rm_rpc_hdr) / sizeof(u32)))
#define RM_RPC_API (RM_RPC_API_VERSION | RM_RPC_HEADER_WORDS)
#define RM_RPC_TYPE_CONTINUATION 0x0
#define RM_RPC_TYPE_REQUEST 0x1
#define RM_RPC_TYPE_REPLY 0x2
#define RM_RPC_TYPE_NOTIF 0x3
#define RM_RPC_TYPE_MASK GENMASK(1, 0)
#define GUNYAH_RM_MAX_NUM_FRAGMENTS 62
#define RM_RPC_FRAGMENTS_MASK GENMASK(7, 2)
/* clang-format on */
struct gunyah_rm_rpc_hdr {
u8 api;
u8 type;
__le16 seq;
__le32 msg_id;
} __packed;
struct gunyah_rm_rpc_reply_hdr {
struct gunyah_rm_rpc_hdr hdr;
__le32 err_code; /* GUNYAH_RM_ERROR_* */
} __packed;
#define GUNYAH_RM_MSGQ_MSG_SIZE 240
#define GUNYAH_RM_PAYLOAD_SIZE \
(GUNYAH_RM_MSGQ_MSG_SIZE - sizeof(struct gunyah_rm_rpc_hdr))
/* RM Error codes */
enum gunyah_rm_error {
/* clang-format off */
GUNYAH_RM_ERROR_OK = 0x0,
GUNYAH_RM_ERROR_UNIMPLEMENTED = 0xFFFFFFFF,
GUNYAH_RM_ERROR_NOMEM = 0x1,
GUNYAH_RM_ERROR_NORESOURCE = 0x2,
GUNYAH_RM_ERROR_DENIED = 0x3,
GUNYAH_RM_ERROR_INVALID = 0x4,
GUNYAH_RM_ERROR_BUSY = 0x5,
GUNYAH_RM_ERROR_ARGUMENT_INVALID = 0x6,
GUNYAH_RM_ERROR_HANDLE_INVALID = 0x7,
GUNYAH_RM_ERROR_VALIDATE_FAILED = 0x8,
GUNYAH_RM_ERROR_MAP_FAILED = 0x9,
GUNYAH_RM_ERROR_MEM_INVALID = 0xA,
GUNYAH_RM_ERROR_MEM_INUSE = 0xB,
GUNYAH_RM_ERROR_MEM_RELEASED = 0xC,
GUNYAH_RM_ERROR_VMID_INVALID = 0xD,
GUNYAH_RM_ERROR_LOOKUP_FAILED = 0xE,
GUNYAH_RM_ERROR_IRQ_INVALID = 0xF,
GUNYAH_RM_ERROR_IRQ_INUSE = 0x10,
GUNYAH_RM_ERROR_IRQ_RELEASED = 0x11,
/* clang-format on */
};
/**
* struct gunyah_rm_message - Represents a complete message from resource manager
* @payload: Combined payload of all the fragments (msg headers stripped off).
* @size: Size of the payload received so far.
* @msg_id: Message ID from the header.
* @type: RM_RPC_TYPE_REPLY or RM_RPC_TYPE_NOTIF.
* @num_fragments: total number of fragments expected to be received.
* @fragments_received: fragments received so far.
* @reply: Fields used for request/reply sequences
*/
struct gunyah_rm_message {
void *payload;
size_t size;
u32 msg_id;
u8 type;
u8 num_fragments;
u8 fragments_received;
/**
* @ret: Linux return code, there was an error processing message
* @seq: Sequence ID for the main message.
* @rm_error: For request/reply sequences with standard replies
* @seq_done: Signals caller that the RM reply has been received
*/
struct {
int ret;
u16 seq;
enum gunyah_rm_error rm_error;
struct completion seq_done;
} reply;
};
/**
* struct gunyah_rm - private data for communicating w/Gunyah resource manager
* @dev: pointer to RM platform device
* @tx_ghrsc: message queue resource to TX to RM
* @rx_ghrsc: message queue resource to RX from RM
* @active_rx_message: ongoing gunyah_rm_message for which we're receiving fragments
* @call_xarray: xarray to allocate & lookup sequence IDs for Request/Response flows
* @next_seq: next ID to allocate (for xa_alloc_cyclic)
* @recv_msg: cached allocation for Rx messages
* @send_msg: cached allocation for Tx messages. Must hold @send_lock to manipulate.
* @send_lock: synchronization to allow only one request to be sent at a time
* @send_ready: completed when we know Tx message queue can take more messages
* @nh: notifier chain for clients interested in RM notification messages
* @miscdev: /dev/gunyah
* @parent_fwnode: Parent IRQ fwnode to translate Gunyah hwirqs to Linux irqs
*/
struct gunyah_rm {
struct device *dev;
struct gunyah_resource tx_ghrsc;
struct gunyah_resource rx_ghrsc;
struct gunyah_rm_message *active_rx_message;
struct xarray call_xarray;
u32 next_seq;
unsigned char recv_msg[GUNYAH_RM_MSGQ_MSG_SIZE];
unsigned char send_msg[GUNYAH_RM_MSGQ_MSG_SIZE];
struct mutex send_lock;
struct completion send_ready;
struct blocking_notifier_head nh;
struct auxiliary_device adev;
struct miscdevice miscdev;
struct fwnode_handle *parent_fwnode;
};
/**
* gunyah_rm_error_remap() - Remap Gunyah resource manager errors into a Linux error code
* @rm_error: "Standard" return value from Gunyah resource manager
*/
static inline int gunyah_rm_error_remap(enum gunyah_rm_error rm_error)
{
switch (rm_error) {
case GUNYAH_RM_ERROR_OK:
return 0;
case GUNYAH_RM_ERROR_UNIMPLEMENTED:
return -EOPNOTSUPP;
case GUNYAH_RM_ERROR_NOMEM:
return -ENOMEM;
case GUNYAH_RM_ERROR_NORESOURCE:
return -ENODEV;
case GUNYAH_RM_ERROR_DENIED:
return -EPERM;
case GUNYAH_RM_ERROR_BUSY:
return -EBUSY;
case GUNYAH_RM_ERROR_INVALID:
case GUNYAH_RM_ERROR_ARGUMENT_INVALID:
case GUNYAH_RM_ERROR_HANDLE_INVALID:
case GUNYAH_RM_ERROR_VALIDATE_FAILED:
case GUNYAH_RM_ERROR_MAP_FAILED:
case GUNYAH_RM_ERROR_MEM_INVALID:
case GUNYAH_RM_ERROR_MEM_INUSE:
case GUNYAH_RM_ERROR_MEM_RELEASED:
case GUNYAH_RM_ERROR_VMID_INVALID:
case GUNYAH_RM_ERROR_LOOKUP_FAILED:
case GUNYAH_RM_ERROR_IRQ_INVALID:
case GUNYAH_RM_ERROR_IRQ_INUSE:
case GUNYAH_RM_ERROR_IRQ_RELEASED:
return -EINVAL;
default:
return -EBADMSG;
}
}
struct gunyah_resource *
gunyah_rm_alloc_resource(struct gunyah_rm *rm,
struct gunyah_rm_hyp_resource *hyp_resource)
{
struct gunyah_resource *ghrsc;
int ret;
ghrsc = kzalloc(sizeof(*ghrsc), GFP_KERNEL);
if (!ghrsc)
return NULL;
ghrsc->type = hyp_resource->type;
ghrsc->capid = le64_to_cpu(hyp_resource->cap_id);
ghrsc->irq = IRQ_NOTCONNECTED;
ghrsc->rm_label = le32_to_cpu(hyp_resource->resource_label);
if (hyp_resource->virq) {
struct irq_fwspec fwspec;
fwspec.fwnode = rm->parent_fwnode;
ret = arch_gunyah_fill_irq_fwspec_params(le32_to_cpu(hyp_resource->virq), &fwspec);
if (ret) {
dev_err(rm->dev,
"Failed to translate interrupt for resource %d label: %d: %d\n",
ghrsc->type, ghrsc->rm_label, ret);
}
ret = irq_create_fwspec_mapping(&fwspec);
if (ret < 0) {
dev_err(rm->dev,
"Failed to allocate interrupt for resource %d label: %d: %d\n",
ghrsc->type, ghrsc->rm_label, ret);
kfree(ghrsc);
return NULL;
}
ghrsc->irq = ret;
}
return ghrsc;
}
void gunyah_rm_free_resource(struct gunyah_resource *ghrsc)
{
irq_dispose_mapping(ghrsc->irq);
kfree(ghrsc);
}
static int gunyah_rm_init_message_payload(struct gunyah_rm_message *message,
const void *msg, size_t hdr_size,
size_t msg_size)
{
const struct gunyah_rm_rpc_hdr *hdr = msg;
size_t max_buf_size, payload_size;
if (msg_size < hdr_size)
return -EINVAL;
payload_size = msg_size - hdr_size;
message->num_fragments = FIELD_GET(RM_RPC_FRAGMENTS_MASK, hdr->type);
message->fragments_received = 0;
/* There's not going to be any payload, no need to allocate buffer. */
if (!payload_size && !message->num_fragments)
return 0;
if (message->num_fragments > GUNYAH_RM_MAX_NUM_FRAGMENTS)
return -EINVAL;
max_buf_size = payload_size +
(message->num_fragments * GUNYAH_RM_PAYLOAD_SIZE);
message->payload = kzalloc(max_buf_size, GFP_KERNEL);
if (!message->payload)
return -ENOMEM;
memcpy(message->payload, msg + hdr_size, payload_size);
message->size = payload_size;
return 0;
}
static void gunyah_rm_abort_message(struct gunyah_rm *rm)
{
kfree(rm->active_rx_message->payload);
switch (rm->active_rx_message->type) {
case RM_RPC_TYPE_REPLY:
rm->active_rx_message->reply.ret = -EIO;
complete(&rm->active_rx_message->reply.seq_done);
break;
case RM_RPC_TYPE_NOTIF:
fallthrough;
default:
kfree(rm->active_rx_message);
}
rm->active_rx_message = NULL;
}
static inline void gunyah_rm_try_complete_message(struct gunyah_rm *rm)
{
struct gunyah_rm_message *message = rm->active_rx_message;
if (!message || message->fragments_received != message->num_fragments)
return;
switch (message->type) {
case RM_RPC_TYPE_REPLY:
complete(&message->reply.seq_done);
break;
case RM_RPC_TYPE_NOTIF:
blocking_notifier_call_chain(&rm->nh, message->msg_id,
message->payload);
kfree(message->payload);
kfree(message);
break;
default:
dev_err_ratelimited(rm->dev,
"Invalid message type (%u) received\n",
message->type);
gunyah_rm_abort_message(rm);
break;
}
rm->active_rx_message = NULL;
}
static void gunyah_rm_process_notif(struct gunyah_rm *rm, const void *msg,
size_t msg_size)
{
const struct gunyah_rm_rpc_hdr *hdr = msg;
struct gunyah_rm_message *message;
int ret;
if (rm->active_rx_message) {
dev_err(rm->dev,
"Unexpected new notification, still processing an active message");
gunyah_rm_abort_message(rm);
}
message = kzalloc(sizeof(*message), GFP_KERNEL);
if (!message)
return;
message->type = RM_RPC_TYPE_NOTIF;
message->msg_id = le32_to_cpu(hdr->msg_id);
ret = gunyah_rm_init_message_payload(message, msg, sizeof(*hdr),
msg_size);
if (ret) {
dev_err(rm->dev,
"Failed to initialize message for notification: %d\n",
ret);
kfree(message);
return;
}
rm->active_rx_message = message;
gunyah_rm_try_complete_message(rm);
}
static void gunyah_rm_process_reply(struct gunyah_rm *rm, const void *msg,
size_t msg_size)
{
const struct gunyah_rm_rpc_reply_hdr *reply_hdr = msg;
struct gunyah_rm_message *message;
u16 seq_id;
seq_id = le16_to_cpu(reply_hdr->hdr.seq);
message = xa_load(&rm->call_xarray, seq_id);
if (!message || message->msg_id != le32_to_cpu(reply_hdr->hdr.msg_id))
return;
if (rm->active_rx_message) {
dev_err(rm->dev,
"Unexpected new reply, still processing an active message");
gunyah_rm_abort_message(rm);
}
if (gunyah_rm_init_message_payload(message, msg, sizeof(*reply_hdr),
msg_size)) {
dev_err(rm->dev,
"Failed to alloc message buffer for sequence %d\n",
seq_id);
/* Send message complete and error the client. */
message->reply.ret = -ENOMEM;
complete(&message->reply.seq_done);
return;
}
message->reply.rm_error = le32_to_cpu(reply_hdr->err_code);
rm->active_rx_message = message;
gunyah_rm_try_complete_message(rm);
}
static void gunyah_rm_process_cont(struct gunyah_rm *rm,
struct gunyah_rm_message *message,
const void *msg, size_t msg_size)
{
const struct gunyah_rm_rpc_hdr *hdr = msg;
size_t payload_size = msg_size - sizeof(*hdr);
if (!rm->active_rx_message)
return;
/*
* hdr->fragments and hdr->msg_id preserves the value from first reply
* or notif message. To detect mishandling, check it's still intact.
*/
if (message->msg_id != le32_to_cpu(hdr->msg_id) ||
message->num_fragments !=
FIELD_GET(RM_RPC_FRAGMENTS_MASK, hdr->type)) {
gunyah_rm_abort_message(rm);
return;
}
memcpy(message->payload + message->size, msg + sizeof(*hdr),
payload_size);
message->size += payload_size;
message->fragments_received++;
gunyah_rm_try_complete_message(rm);
}
static irqreturn_t gunyah_rm_rx(int irq, void *data)
{
enum gunyah_error gunyah_error;
struct gunyah_rm_rpc_hdr *hdr;
struct gunyah_rm *rm = data;
void *msg = &rm->recv_msg[0];
size_t len;
bool ready;
do {
gunyah_error = gunyah_hypercall_msgq_recv(rm->rx_ghrsc.capid,
msg,
sizeof(rm->recv_msg),
&len, &ready);
if (gunyah_error != GUNYAH_ERROR_OK) {
if (gunyah_error != GUNYAH_ERROR_MSGQUEUE_EMPTY)
dev_warn(rm->dev,
"Failed to receive data: %d\n",
gunyah_error);
return IRQ_HANDLED;
}
if (len < sizeof(*hdr)) {
dev_err_ratelimited(
rm->dev,
"Too small message received. size=%ld\n", len);
continue;
}
hdr = msg;
if (hdr->api != RM_RPC_API) {
dev_err(rm->dev, "Unknown RM RPC API version: %x\n",
hdr->api);
return IRQ_HANDLED;
}
switch (FIELD_GET(RM_RPC_TYPE_MASK, hdr->type)) {
case RM_RPC_TYPE_NOTIF:
gunyah_rm_process_notif(rm, msg, len);
break;
case RM_RPC_TYPE_REPLY:
gunyah_rm_process_reply(rm, msg, len);
break;
case RM_RPC_TYPE_CONTINUATION:
gunyah_rm_process_cont(rm, rm->active_rx_message, msg,
len);
break;
default:
dev_err(rm->dev,
"Invalid message type (%lu) received\n",
FIELD_GET(RM_RPC_TYPE_MASK, hdr->type));
return IRQ_HANDLED;
}
} while (ready);
return IRQ_HANDLED;
}
static irqreturn_t gunyah_rm_tx(int irq, void *data)
{
struct gunyah_rm *rm = data;
complete(&rm->send_ready);
return IRQ_HANDLED;
}
static int gunyah_rm_msgq_send(struct gunyah_rm *rm, size_t size, bool push)
{
const u64 tx_flags = push ? GUNYAH_HYPERCALL_MSGQ_TX_FLAGS_PUSH : 0;
enum gunyah_error gunyah_error;
void *data = &rm->send_msg[0];
bool ready;
lockdep_assert_held(&rm->send_lock);
again:
wait_for_completion(&rm->send_ready);
gunyah_error = gunyah_hypercall_msgq_send(rm->tx_ghrsc.capid, size,
data, tx_flags, &ready);
/* Should never happen because Linux properly tracks the ready-state of the msgq */
if (WARN_ON(gunyah_error == GUNYAH_ERROR_MSGQUEUE_FULL))
goto again;
if (ready)
complete(&rm->send_ready);
return gunyah_error_remap(gunyah_error);
}
static int gunyah_rm_send_request(struct gunyah_rm *rm, u32 message_id,
const void *req_buf, size_t req_buf_size,
struct gunyah_rm_message *message)
{
size_t buf_size_remaining = req_buf_size;
const void *req_buf_curr = req_buf;
struct gunyah_rm_rpc_hdr *hdr =
(struct gunyah_rm_rpc_hdr *)&rm->send_msg[0];
struct gunyah_rm_rpc_hdr hdr_template;
void *payload = hdr + 1;
u32 cont_fragments = 0;
size_t payload_size;
bool push;
int ret;
if (req_buf_size >
GUNYAH_RM_MAX_NUM_FRAGMENTS * GUNYAH_RM_PAYLOAD_SIZE) {
dev_warn(
rm->dev,
"Limit (%lu bytes) exceeded for the maximum message size: %lu\n",
GUNYAH_RM_MAX_NUM_FRAGMENTS * GUNYAH_RM_PAYLOAD_SIZE,
req_buf_size);
dump_stack();
return -E2BIG;
}
if (req_buf_size)
cont_fragments = (req_buf_size - 1) / GUNYAH_RM_PAYLOAD_SIZE;
hdr_template.api = RM_RPC_API;
hdr_template.type = FIELD_PREP(RM_RPC_TYPE_MASK, RM_RPC_TYPE_REQUEST) |
FIELD_PREP(RM_RPC_FRAGMENTS_MASK, cont_fragments);
hdr_template.seq = cpu_to_le16(message->reply.seq);
hdr_template.msg_id = cpu_to_le32(message_id);
do {
*hdr = hdr_template;
/* Copy payload */
payload_size = min(buf_size_remaining, GUNYAH_RM_PAYLOAD_SIZE);
memcpy(payload, req_buf_curr, payload_size);
req_buf_curr += payload_size;
buf_size_remaining -= payload_size;
/* Only the last message should have push flag set */
push = !buf_size_remaining;
ret = gunyah_rm_msgq_send(rm, sizeof(*hdr) + payload_size,
push);
if (ret)
break;
hdr_template.type =
FIELD_PREP(RM_RPC_TYPE_MASK, RM_RPC_TYPE_CONTINUATION) |
FIELD_PREP(RM_RPC_FRAGMENTS_MASK, cont_fragments);
} while (buf_size_remaining);
return ret;
}
/**
* gunyah_rm_call: Achieve request-response type communication with RPC
* @rm: Pointer to Gunyah resource manager internal data
* @message_id: The RM RPC message-id
* @req_buf: Request buffer that contains the payload
* @req_buf_size: Total size of the payload
* @resp_buf: Pointer to a response buffer
* @resp_buf_size: Size of the response buffer
*
* Make a request to the Resource Manager and wait for reply back. For a successful
* response, the function returns the payload. The size of the payload is set in
* resp_buf_size. The resp_buf must be freed by the caller when 0 is returned
* and resp_buf_size != 0.
*
* req_buf should be not NULL for req_buf_size >0. If req_buf_size == 0,
* req_buf *can* be NULL and no additional payload is sent.
*
* Context: Process context. Will sleep waiting for reply.
* Return: 0 on success. <0 if error.
*/
int gunyah_rm_call(struct gunyah_rm *rm, u32 message_id, const void *req_buf,
size_t req_buf_size, void **resp_buf, size_t *resp_buf_size)
{
struct gunyah_rm_message message = { 0 };
u32 seq_id;
int ret;
/* message_id 0 is reserved. req_buf_size implies req_buf is not NULL */
if (!rm || !message_id || (!req_buf && req_buf_size))
return -EINVAL;
message.type = RM_RPC_TYPE_REPLY;
message.msg_id = message_id;
message.reply.seq_done =
COMPLETION_INITIALIZER_ONSTACK(message.reply.seq_done);
/* Allocate a new seq number for this message */
ret = xa_alloc_cyclic(&rm->call_xarray, &seq_id, &message, xa_limit_16b,
&rm->next_seq, GFP_KERNEL);
if (ret < 0)
return ret;
message.reply.seq = lower_16_bits(seq_id);
mutex_lock(&rm->send_lock);
/* Send the request to the Resource Manager */
ret = gunyah_rm_send_request(rm, message_id, req_buf, req_buf_size,
&message);
if (ret < 0) {
dev_warn(rm->dev, "Failed to send request. Error: %d\n", ret);
goto out;
}
/*
* Wait for response. Uninterruptible because rollback based on what RM did to VM
* requires us to know how RM handled the call.
*/
wait_for_completion(&message.reply.seq_done);
/* Check for internal (kernel) error waiting for the response */
if (message.reply.ret) {
ret = message.reply.ret;
goto out;
}
/* Got a response, did resource manager give us an error? */
if (message.reply.rm_error != GUNYAH_RM_ERROR_OK) {
dev_warn(rm->dev, "RM rejected message %08x. Error: %d\n",
message_id, message.reply.rm_error);
ret = gunyah_rm_error_remap(message.reply.rm_error);
kfree(message.payload);
goto out;
}
/* Everything looks good, return the payload */
if (resp_buf_size)
*resp_buf_size = message.size;
if (message.size && resp_buf) {
*resp_buf = message.payload;
} else {
/* kfree in case RM sent us multiple fragments but never any data in
* those fragments. We would've allocated memory for it, but message.size == 0
*/
kfree(message.payload);
}
out:
mutex_unlock(&rm->send_lock);
xa_erase(&rm->call_xarray, message.reply.seq);
return ret;
}
EXPORT_SYMBOL_GPL(gunyah_rm_call);
int gunyah_rm_notifier_register(struct gunyah_rm *rm, struct notifier_block *nb)
{
return blocking_notifier_chain_register(&rm->nh, nb);
}
EXPORT_SYMBOL_GPL(gunyah_rm_notifier_register);
int gunyah_rm_notifier_unregister(struct gunyah_rm *rm,
struct notifier_block *nb)
{
return blocking_notifier_chain_unregister(&rm->nh, nb);
}
EXPORT_SYMBOL_GPL(gunyah_rm_notifier_unregister);
struct device *gunyah_rm_get(struct gunyah_rm *rm)
{
return get_device(rm->miscdev.this_device);
}
EXPORT_SYMBOL_GPL(gunyah_rm_get);
void gunyah_rm_put(struct gunyah_rm *rm)
{
put_device(rm->miscdev.this_device);
}
EXPORT_SYMBOL_GPL(gunyah_rm_put);
static long gunyah_dev_ioctl(struct file *filp, unsigned int cmd,
unsigned long arg)
{
struct miscdevice *miscdev = filp->private_data;
struct gunyah_rm *rm = container_of(miscdev, struct gunyah_rm, miscdev);
return gunyah_dev_vm_mgr_ioctl(rm, cmd, arg);
}
static const struct file_operations gunyah_dev_fops = {
/* clang-format off */
.owner = THIS_MODULE,
.unlocked_ioctl = gunyah_dev_ioctl,
.compat_ioctl = compat_ptr_ioctl,
.llseek = noop_llseek,
/* clang-format on */
};
static int gunyah_platform_probe_capability(struct platform_device *pdev,
int idx,
struct gunyah_resource *ghrsc)
{
int ret;
ghrsc->irq = platform_get_irq(pdev, idx);
if (ghrsc->irq < 0) {
dev_err(&pdev->dev, "Failed to get %s irq: %d\n",
idx ? "rx" : "tx", ghrsc->irq);
return ghrsc->irq;
}
ret = of_property_read_u64_index(pdev->dev.of_node, "reg", idx,
&ghrsc->capid);
if (ret) {
dev_err(&pdev->dev, "Failed to get %s capid: %d\n",
idx ? "rx" : "tx", ret);
return ret;
}
return 0;
}
static int gunyah_rm_probe_tx_msgq(struct gunyah_rm *rm,
struct platform_device *pdev)
{
int ret;
rm->tx_ghrsc.type = GUNYAH_RESOURCE_TYPE_MSGQ_TX;
ret = gunyah_platform_probe_capability(pdev, 0, &rm->tx_ghrsc);
if (ret)
return ret;
enable_irq_wake(rm->tx_ghrsc.irq);
return devm_request_irq(rm->dev, rm->tx_ghrsc.irq, gunyah_rm_tx, 0,
"gunyah_rm_tx", rm);
}
static int gunyah_rm_probe_rx_msgq(struct gunyah_rm *rm,
struct platform_device *pdev)
{
int ret;
rm->rx_ghrsc.type = GUNYAH_RESOURCE_TYPE_MSGQ_RX;
ret = gunyah_platform_probe_capability(pdev, 1, &rm->rx_ghrsc);
if (ret)
return ret;
enable_irq_wake(rm->rx_ghrsc.irq);
return devm_request_threaded_irq(rm->dev, rm->rx_ghrsc.irq, NULL,
gunyah_rm_rx, IRQF_ONESHOT,
"gunyah_rm_rx", rm);
}
static void gunyah_adev_release(struct device *dev)
{
/* no-op */
}
static int gunyah_adev_init(struct gunyah_rm *rm, const char *name)
{
struct auxiliary_device *adev = &rm->adev;
int ret = 0;
adev->name = name;
adev->dev.parent = rm->dev;
adev->dev.release = gunyah_adev_release;
ret = auxiliary_device_init(adev);
if (ret)
return ret;
ret = auxiliary_device_add(adev);
if (ret) {
auxiliary_device_uninit(adev);
return ret;
}
return ret;
}
static int gunyah_rm_probe(struct platform_device *pdev)
{
struct device_node *parent_irq_node;
struct gunyah_rm *rm;
int ret;
rm = devm_kzalloc(&pdev->dev, sizeof(*rm), GFP_KERNEL);
if (!rm)
return -ENOMEM;
platform_set_drvdata(pdev, rm);
rm->dev = &pdev->dev;
mutex_init(&rm->send_lock);
init_completion(&rm->send_ready);
BLOCKING_INIT_NOTIFIER_HEAD(&rm->nh);
xa_init_flags(&rm->call_xarray, XA_FLAGS_ALLOC);
device_init_wakeup(&pdev->dev, true);
ret = gunyah_rm_probe_tx_msgq(rm, pdev);
if (ret)
return ret;
/* assume RM is ready to receive messages from us */
complete(&rm->send_ready);
ret = gunyah_rm_probe_rx_msgq(rm, pdev);
if (ret)
return ret;
parent_irq_node = of_irq_find_parent(pdev->dev.of_node);
if (!parent_irq_node) {
dev_err(&pdev->dev,
"Failed to find interrupt parent of resource manager\n");
return -ENODEV;
}
rm->parent_fwnode = of_node_to_fwnode(parent_irq_node);
if (!rm->parent_fwnode) {
dev_err(&pdev->dev,
"Failed to find interrupt parent domain of resource manager\n");
return -ENODEV;
}
rm->miscdev.parent = &pdev->dev;
rm->miscdev.name = "gunyah";
rm->miscdev.minor = MISC_DYNAMIC_MINOR;
rm->miscdev.fops = &gunyah_dev_fops;
ret = misc_register(&rm->miscdev);
if (ret) {
dev_err(&pdev->dev, "Failed to register gunyah misc device\n");
} else {
ret = gunyah_adev_init(rm, "gh_rm_core");
if (ret) {
dev_err(&pdev->dev, "Failed to add gh_rm_core device\n");
return ret;
}
}
return 0;
}
static void gunyah_rm_remove(struct platform_device *pdev)
{
struct gunyah_rm *rm = platform_get_drvdata(pdev);
misc_deregister(&rm->miscdev);
}
static const struct of_device_id gunyah_rm_of_match[] = {
{ .compatible = "gunyah-resource-manager" },
{}
};
MODULE_DEVICE_TABLE(of, gunyah_rm_of_match);
static struct platform_driver gunyah_rm_driver = {
.probe = gunyah_rm_probe,
.remove_new = gunyah_rm_remove,
.driver = {
.name = "gunyah_rsc_mgr",
.of_match_table = gunyah_rm_of_match,
},
};
module_platform_driver(gunyah_rm_driver);
MODULE_LICENSE("GPL");
MODULE_DESCRIPTION("Gunyah Resource Manager Driver");