diff options
Diffstat (limited to 'libmctp/astlpc.c')
-rw-r--r-- | libmctp/astlpc.c | 1497 |
1 files changed, 1497 insertions, 0 deletions
diff --git a/libmctp/astlpc.c b/libmctp/astlpc.c new file mode 100644 index 0000000..dbd1f0e --- /dev/null +++ b/libmctp/astlpc.c @@ -0,0 +1,1497 @@ +/* SPDX-License-Identifier: Apache-2.0 OR GPL-2.0-or-later */ + +#if HAVE_CONFIG_H +#include "config.h" +#endif + +#if HAVE_ENDIAN_H +#include <endian.h> +#endif + +#include <assert.h> +#include <errno.h> +#include <inttypes.h> +#include <stdbool.h> +#include <stdlib.h> +#include <string.h> + +#define pr_fmt(x) "astlpc: " x + +#include "container_of.h" +#include "crc32.h" +#include "libmctp.h" +#include "libmctp-alloc.h" +#include "libmctp-log.h" +#include "libmctp-astlpc.h" +#include "range.h" + +#ifdef MCTP_HAVE_FILEIO + +#include <unistd.h> +#include <fcntl.h> +#include <poll.h> +#include <sys/ioctl.h> +#include <sys/mman.h> +#include <linux/aspeed-lpc-ctrl.h> + +/* kernel interface */ +static const char *kcs_path = "/dev/mctp0"; +static const char *lpc_path = "/dev/aspeed-lpc-ctrl"; + +#endif + +enum mctp_astlpc_cmd { + cmd_initialise = 0x00, + cmd_tx_begin = 0x01, + cmd_rx_complete = 0x02, + cmd_dummy_value = 0xff, +}; + +enum mctp_astlpc_buffer_state { + /* + * Prior to "Channel Ready" we mark the buffers as "idle" to catch illegal accesses. In this + * state neither side is considered the owner of the buffer. + * + * Upon "Channel Ready", each side transitions the buffers from the initial "idle" state + * to the following target states: + * + * Tx buffer: "acquired" + * Rx buffer: "released" + */ + buffer_state_idle, + + /* + * Beyond initialisation by "Channel Ready", buffers are in the "acquired" state once: + * + * 1. We dequeue a control command transferring the buffer to our ownership out of the KCS + * interface, and + * 2. We are yet to complete all of our required accesses to the buffer + * + * * The Tx buffer enters the "acquired" state when we dequeue the "Rx Complete" command + * * The Rx buffer enters the "acquired" state when we dequeue the "Tx Begin" command + * + * It is a failure of implementation if it's possible for both sides to simultaneously + * consider a buffer as "acquired". + */ + buffer_state_acquired, + + /* + * Buffers are in the "prepared" state when: + * + * 1. We have completed all of our required accesses (read or write) for the buffer, and + * 2. We have not yet successfully enqueued the control command to hand off ownership + */ + buffer_state_prepared, + + /* + * Beyond initialisation by "Channel Ready", buffers are in the "released" state once: + * + * 1. We successfully enqueue the control command transferring ownership to the remote + * side in to the KCS interface + * + * * The Tx buffer enters the "released" state when we enqueue the "Tx Begin" command + * * The Rx buffer enters the "released" state when we enqueue the "Rx Complete" command + * + * It may be the case that both sides simultaneously consider a buffer to be in the + * "released" state. However, if this is true, it must also be true that a buffer ownership + * transfer command has been enqueued in the KCS interface and is yet to be dequeued. + */ + buffer_state_released, +}; + +struct mctp_astlpc_buffer { + uint32_t offset; + uint32_t size; + enum mctp_astlpc_buffer_state state; +}; + +struct mctp_astlpc_layout { + struct mctp_astlpc_buffer rx; + struct mctp_astlpc_buffer tx; +}; + +struct mctp_astlpc_protocol { + uint16_t version; + uint32_t (*packet_size)(uint32_t body); + uint32_t (*body_size)(uint32_t packet); + void (*pktbuf_protect)(struct mctp_pktbuf *pkt); + bool (*pktbuf_validate)(struct mctp_pktbuf *pkt); +}; + +struct mctp_binding_astlpc { + struct mctp_binding binding; + + void *lpc_map; + struct mctp_astlpc_layout layout; + + uint8_t mode; + uint32_t requested_mtu; + + const struct mctp_astlpc_protocol *proto; + + /* direct ops data */ + struct mctp_binding_astlpc_ops ops; + void *ops_data; + + /* fileio ops data */ + int kcs_fd; + uint8_t kcs_status; +}; + +#define binding_to_astlpc(b) \ + container_of(b, struct mctp_binding_astlpc, binding) + +#define astlpc_prlog(ctx, lvl, fmt, ...) \ + do { \ + bool __bmc = ((ctx)->mode == MCTP_BINDING_ASTLPC_MODE_BMC); \ + mctp_prlog(lvl, pr_fmt("%s: " fmt), __bmc ? "bmc" : "host", \ + ##__VA_ARGS__); \ + } while (0) + +#define astlpc_prerr(ctx, fmt, ...) \ + astlpc_prlog(ctx, MCTP_LOG_ERR, fmt, ##__VA_ARGS__) +#define astlpc_prwarn(ctx, fmt, ...) \ + astlpc_prlog(ctx, MCTP_LOG_WARNING, fmt, ##__VA_ARGS__) +#define astlpc_prinfo(ctx, fmt, ...) \ + astlpc_prlog(ctx, MCTP_LOG_INFO, fmt, ##__VA_ARGS__) +#define astlpc_prdebug(ctx, fmt, ...) \ + astlpc_prlog(ctx, MCTP_LOG_DEBUG, fmt, ##__VA_ARGS__) + +/* clang-format off */ +#define ASTLPC_MCTP_MAGIC 0x4d435450 +#define ASTLPC_VER_BAD 0 +#define ASTLPC_VER_MIN 1 + +/* Support testing of new binding protocols */ +#ifndef ASTLPC_VER_CUR +#define ASTLPC_VER_CUR 3 +#endif +/* clang-format on */ + +#ifndef ARRAY_SIZE +#define ARRAY_SIZE(a) (sizeof(a) / sizeof(a[0])) +#endif + +static uint32_t astlpc_packet_size_v1(uint32_t body) +{ + assert((body + 4) > body); + + return body + 4; +} + +static uint32_t astlpc_body_size_v1(uint32_t packet) +{ + assert((packet - 4) < packet); + + return packet - 4; +} + +void astlpc_pktbuf_protect_v1(struct mctp_pktbuf *pkt) +{ + (void)pkt; +} + +bool astlpc_pktbuf_validate_v1(struct mctp_pktbuf *pkt) +{ + (void)pkt; + return true; +} + +static uint32_t astlpc_packet_size_v3(uint32_t body) +{ + assert((body + 4 + 4) > body); + + return body + 4 + 4; +} + +static uint32_t astlpc_body_size_v3(uint32_t packet) +{ + assert((packet - 4 - 4) < packet); + + return packet - 4 - 4; +} + +void astlpc_pktbuf_protect_v3(struct mctp_pktbuf *pkt) +{ + uint32_t code; + + code = htobe32(crc32(mctp_pktbuf_hdr(pkt), mctp_pktbuf_size(pkt))); + mctp_prdebug("%s: 0x%" PRIx32, __func__, code); + mctp_pktbuf_push(pkt, &code, 4); +} + +bool astlpc_pktbuf_validate_v3(struct mctp_pktbuf *pkt) +{ + uint32_t code; + void *check; + + code = be32toh(crc32(mctp_pktbuf_hdr(pkt), mctp_pktbuf_size(pkt) - 4)); + mctp_prdebug("%s: 0x%" PRIx32, __func__, code); + check = mctp_pktbuf_pop(pkt, 4); + return check && !memcmp(&code, check, 4); +} + +static const struct mctp_astlpc_protocol astlpc_protocol_version[] = { + [0] = { + .version = 0, + .packet_size = NULL, + .body_size = NULL, + .pktbuf_protect = NULL, + .pktbuf_validate = NULL, + }, + [1] = { + .version = 1, + .packet_size = astlpc_packet_size_v1, + .body_size = astlpc_body_size_v1, + .pktbuf_protect = astlpc_pktbuf_protect_v1, + .pktbuf_validate = astlpc_pktbuf_validate_v1, + }, + [2] = { + .version = 2, + .packet_size = astlpc_packet_size_v1, + .body_size = astlpc_body_size_v1, + .pktbuf_protect = astlpc_pktbuf_protect_v1, + .pktbuf_validate = astlpc_pktbuf_validate_v1, + }, + [3] = { + .version = 3, + .packet_size = astlpc_packet_size_v3, + .body_size = astlpc_body_size_v3, + .pktbuf_protect = astlpc_pktbuf_protect_v3, + .pktbuf_validate = astlpc_pktbuf_validate_v3, + }, +}; + +struct mctp_lpcmap_hdr { + uint32_t magic; + + uint16_t bmc_ver_min; + uint16_t bmc_ver_cur; + uint16_t host_ver_min; + uint16_t host_ver_cur; + uint16_t negotiated_ver; + uint16_t pad0; + + struct { + uint32_t rx_offset; + uint32_t rx_size; + uint32_t tx_offset; + uint32_t tx_size; + } layout; +} __attribute__((packed)); + +static const uint32_t control_size = 0x100; + +#define LPC_WIN_SIZE (1 * 1024 * 1024) + +#define KCS_STATUS_BMC_READY 0x80 +#define KCS_STATUS_CHANNEL_ACTIVE 0x40 +#define KCS_STATUS_IBF 0x02 +#define KCS_STATUS_OBF 0x01 + +static inline int mctp_astlpc_kcs_write(struct mctp_binding_astlpc *astlpc, + enum mctp_binding_astlpc_kcs_reg reg, + uint8_t val) +{ + return astlpc->ops.kcs_write(astlpc->ops_data, reg, val); +} + +static inline int mctp_astlpc_kcs_read(struct mctp_binding_astlpc *astlpc, + enum mctp_binding_astlpc_kcs_reg reg, + uint8_t *val) +{ + return astlpc->ops.kcs_read(astlpc->ops_data, reg, val); +} + +static inline int mctp_astlpc_lpc_write(struct mctp_binding_astlpc *astlpc, + const void *buf, long offset, + size_t len) +{ + astlpc_prdebug(astlpc, "%s: %zu bytes to 0x%lx", __func__, len, offset); + + assert(offset >= 0); + + /* Indirect access */ + if (astlpc->ops.lpc_write) { + void *data = astlpc->ops_data; + + return astlpc->ops.lpc_write(data, buf, offset, len); + } + + /* Direct mapping */ + assert(astlpc->lpc_map); + memcpy(&((char *)astlpc->lpc_map)[offset], buf, len); + + return 0; +} + +static inline int mctp_astlpc_lpc_read(struct mctp_binding_astlpc *astlpc, + void *buf, long offset, size_t len) +{ + astlpc_prdebug(astlpc, "%s: %zu bytes from 0x%lx", __func__, len, + offset); + + assert(offset >= 0); + + /* Indirect access */ + if (astlpc->ops.lpc_read) { + void *data = astlpc->ops_data; + + return astlpc->ops.lpc_read(data, buf, offset, len); + } + + /* Direct mapping */ + assert(astlpc->lpc_map); + memcpy(buf, &((char *)astlpc->lpc_map)[offset], len); + + return 0; +} + +static int mctp_astlpc_kcs_set_status(struct mctp_binding_astlpc *astlpc, + uint8_t status) +{ + uint8_t data; + int rc; + + /* Since we're setting the status register, we want the other endpoint + * to be interrupted. However, some hardware may only raise a host-side + * interrupt on an ODR event. + * So, write a dummy value of 0xff to ODR, which will ensure that an + * interrupt is triggered, and can be ignored by the host. + */ + data = cmd_dummy_value; + + rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_STATUS, status); + if (rc) { + astlpc_prwarn(astlpc, "KCS status write failed"); + return -1; + } + + rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_DATA, data); + if (rc) { + astlpc_prwarn(astlpc, "KCS dummy data write failed"); + return -1; + } + + return 0; +} + +static int mctp_astlpc_layout_read(struct mctp_binding_astlpc *astlpc, + struct mctp_astlpc_layout *layout) +{ + struct mctp_lpcmap_hdr hdr; + int rc; + + rc = mctp_astlpc_lpc_read(astlpc, &hdr, 0, sizeof(hdr)); + if (rc < 0) + return rc; + + /* Flip the buffers as the names are defined in terms of the host */ + if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC) { + layout->rx.offset = be32toh(hdr.layout.tx_offset); + layout->rx.size = be32toh(hdr.layout.tx_size); + layout->tx.offset = be32toh(hdr.layout.rx_offset); + layout->tx.size = be32toh(hdr.layout.rx_size); + } else { + assert(astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST); + + layout->rx.offset = be32toh(hdr.layout.rx_offset); + layout->rx.size = be32toh(hdr.layout.rx_size); + layout->tx.offset = be32toh(hdr.layout.tx_offset); + layout->tx.size = be32toh(hdr.layout.tx_size); + } + + return 0; +} + +static int mctp_astlpc_layout_write(struct mctp_binding_astlpc *astlpc, + struct mctp_astlpc_layout *layout) +{ + uint32_t rx_size_be; + + if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC) { + struct mctp_lpcmap_hdr hdr; + + /* + * Flip the buffers as the names are defined in terms of the + * host + */ + hdr.layout.rx_offset = htobe32(layout->tx.offset); + hdr.layout.rx_size = htobe32(layout->tx.size); + hdr.layout.tx_offset = htobe32(layout->rx.offset); + hdr.layout.tx_size = htobe32(layout->rx.size); + + return mctp_astlpc_lpc_write(astlpc, &hdr.layout, + offsetof(struct mctp_lpcmap_hdr, + layout), + sizeof(hdr.layout)); + } + + assert(astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST); + + /* + * As of v2 we only need to write rx_size - the offsets are controlled + * by the BMC, as is the BMC's rx_size (host tx_size). + */ + rx_size_be = htobe32(layout->rx.size); + return mctp_astlpc_lpc_write(astlpc, &rx_size_be, + offsetof(struct mctp_lpcmap_hdr, + layout.rx_size), + sizeof(rx_size_be)); +} + +static bool +mctp_astlpc_buffer_validate(const struct mctp_binding_astlpc *astlpc, + const struct mctp_astlpc_buffer *buf, + const char *name) +{ + /* Check for overflow */ + if (buf->offset + buf->size < buf->offset) { + mctp_prerr( + "%s packet buffer parameters overflow: offset: 0x%" PRIx32 + ", size: %" PRIu32, + name, buf->offset, buf->size); + return false; + } + + /* Check that the buffers are contained within the allocated space */ + if (buf->offset + buf->size > LPC_WIN_SIZE) { + mctp_prerr( + "%s packet buffer parameters exceed %uM window size: offset: 0x%" PRIx32 + ", size: %" PRIu32, + name, (LPC_WIN_SIZE / (1024 * 1024)), buf->offset, + buf->size); + return false; + } + + /* Check that the baseline transmission unit is supported */ + if (buf->size < + astlpc->proto->packet_size(MCTP_PACKET_SIZE(MCTP_BTU))) { + mctp_prerr( + "%s packet buffer too small: Require %" PRIu32 + " bytes to support the %u byte baseline transmission unit, found %" PRIu32, + name, + astlpc->proto->packet_size(MCTP_PACKET_SIZE(MCTP_BTU)), + MCTP_BTU, buf->size); + return false; + } + + /* Check for overlap with the control space */ + if (buf->offset < control_size) { + mctp_prerr( + "%s packet buffer overlaps control region {0x%" PRIx32 + ", %" PRIu32 "}: Rx {0x%" PRIx32 ", %" PRIu32 "}", + name, 0U, control_size, buf->offset, buf->size); + return false; + } + + return true; +} + +static bool +mctp_astlpc_layout_validate(const struct mctp_binding_astlpc *astlpc, + const struct mctp_astlpc_layout *layout) +{ + const struct mctp_astlpc_buffer *rx = &layout->rx; + const struct mctp_astlpc_buffer *tx = &layout->tx; + bool rx_valid, tx_valid; + + rx_valid = mctp_astlpc_buffer_validate(astlpc, rx, "Rx"); + tx_valid = mctp_astlpc_buffer_validate(astlpc, tx, "Tx"); + + if (!(rx_valid && tx_valid)) + return false; + + /* Check that the buffers are disjoint */ + if ((rx->offset <= tx->offset && rx->offset + rx->size > tx->offset) || + (tx->offset <= rx->offset && tx->offset + tx->size > rx->offset)) { + mctp_prerr("Rx and Tx packet buffers overlap: Rx {0x%" PRIx32 + ", %" PRIu32 "}, Tx {0x%" PRIx32 ", %" PRIu32 "}", + rx->offset, rx->size, tx->offset, tx->size); + return false; + } + + return true; +} + +static int mctp_astlpc_init_bmc(struct mctp_binding_astlpc *astlpc) +{ + struct mctp_lpcmap_hdr hdr = { 0 }; + uint8_t status; + uint32_t sz; + + /* + * The largest buffer size is half of the allocated MCTP space + * excluding the control space. + */ + sz = ((LPC_WIN_SIZE - control_size) / 2); + + /* + * Trim the MTU to a multiple of 16 to meet the requirements of 12.17 + * Query Hop in DSP0236 v1.3.0. + */ + sz = MCTP_BODY_SIZE(astlpc->proto->body_size(sz)); + sz &= ~0xfUL; + sz = astlpc->proto->packet_size(MCTP_PACKET_SIZE(sz)); + + if (astlpc->requested_mtu) { + uint32_t rpkt, rmtu; + + rmtu = astlpc->requested_mtu; + rpkt = astlpc->proto->packet_size(MCTP_PACKET_SIZE(rmtu)); + sz = MIN(sz, rpkt); + } + + /* Flip the buffers as the names are defined in terms of the host */ + astlpc->layout.tx.offset = control_size; + astlpc->layout.tx.size = sz; + astlpc->layout.rx.offset = + astlpc->layout.tx.offset + astlpc->layout.tx.size; + astlpc->layout.rx.size = sz; + + if (!mctp_astlpc_layout_validate(astlpc, &astlpc->layout)) { + astlpc_prerr(astlpc, "Cannot support an MTU of %" PRIu32, sz); + return -EINVAL; + } + + hdr = (struct mctp_lpcmap_hdr){ + .magic = htobe32(ASTLPC_MCTP_MAGIC), + .bmc_ver_min = htobe16(ASTLPC_VER_MIN), + .bmc_ver_cur = htobe16(ASTLPC_VER_CUR), + + /* Flip the buffers back as we're now describing the host's + * configuration to the host */ + .layout.rx_offset = htobe32(astlpc->layout.tx.offset), + .layout.rx_size = htobe32(astlpc->layout.tx.size), + .layout.tx_offset = htobe32(astlpc->layout.rx.offset), + .layout.tx_size = htobe32(astlpc->layout.rx.size), + }; + + mctp_astlpc_lpc_write(astlpc, &hdr, 0, sizeof(hdr)); + + /* + * Set status indicating that the BMC is now active. Be explicit about + * clearing OBF; we're reinitialising the binding and so any previous + * buffer state is irrelevant. + */ + status = KCS_STATUS_BMC_READY & ~KCS_STATUS_OBF; + return mctp_astlpc_kcs_set_status(astlpc, status); +} + +static int mctp_binding_astlpc_start_bmc(struct mctp_binding *b) +{ + struct mctp_binding_astlpc *astlpc = + container_of(b, struct mctp_binding_astlpc, binding); + + astlpc->proto = &astlpc_protocol_version[ASTLPC_VER_CUR]; + + return mctp_astlpc_init_bmc(astlpc); +} + +static bool mctp_astlpc_validate_version(uint16_t bmc_ver_min, + uint16_t bmc_ver_cur, + uint16_t host_ver_min, + uint16_t host_ver_cur) +{ + if (!(bmc_ver_min && bmc_ver_cur && host_ver_min && host_ver_cur)) { + mctp_prerr("Invalid version present in [%" PRIu16 ", %" PRIu16 + "], [%" PRIu16 ", %" PRIu16 "]", + bmc_ver_min, bmc_ver_cur, host_ver_min, + host_ver_cur); + return false; + } else if (bmc_ver_min > bmc_ver_cur) { + mctp_prerr("Invalid bmc version range [%" PRIu16 ", %" PRIu16 + "]", + bmc_ver_min, bmc_ver_cur); + return false; + } else if (host_ver_min > host_ver_cur) { + mctp_prerr("Invalid host version range [%" PRIu16 ", %" PRIu16 + "]", + host_ver_min, host_ver_cur); + return false; + } else if ((host_ver_cur < bmc_ver_min) || + (host_ver_min > bmc_ver_cur)) { + mctp_prerr( + "Unable to satisfy version negotiation with ranges [%" PRIu16 + ", %" PRIu16 "] and [%" PRIu16 ", %" PRIu16 "]", + bmc_ver_min, bmc_ver_cur, host_ver_min, host_ver_cur); + return false; + } + + return true; +} + +static int mctp_astlpc_negotiate_layout_host(struct mctp_binding_astlpc *astlpc) +{ + struct mctp_astlpc_layout layout; + uint32_t rmtu; + uint32_t sz; + int rc; + + rc = mctp_astlpc_layout_read(astlpc, &layout); + if (rc < 0) + return rc; + + if (!mctp_astlpc_layout_validate(astlpc, &layout)) { + astlpc_prerr( + astlpc, + "BMC provided invalid buffer layout: Rx {0x%" PRIx32 + ", %" PRIu32 "}, Tx {0x%" PRIx32 ", %" PRIu32 "}", + layout.rx.offset, layout.rx.size, layout.tx.offset, + layout.tx.size); + return -EINVAL; + } + + astlpc_prinfo(astlpc, "Desire an MTU of %" PRIu32 " bytes", + astlpc->requested_mtu); + + rmtu = astlpc->requested_mtu; + sz = astlpc->proto->packet_size(MCTP_PACKET_SIZE(rmtu)); + layout.rx.size = sz; + + if (!mctp_astlpc_layout_validate(astlpc, &layout)) { + astlpc_prerr( + astlpc, + "Generated invalid buffer layout with size %" PRIu32 + ": Rx {0x%" PRIx32 ", %" PRIu32 "}, Tx {0x%" PRIx32 + ", %" PRIu32 "}", + sz, layout.rx.offset, layout.rx.size, layout.tx.offset, + layout.tx.size); + return -EINVAL; + } + + astlpc_prinfo(astlpc, "Requesting MTU of %" PRIu32 " bytes", + astlpc->requested_mtu); + + return mctp_astlpc_layout_write(astlpc, &layout); +} + +static uint16_t mctp_astlpc_negotiate_version(uint16_t bmc_ver_min, + uint16_t bmc_ver_cur, + uint16_t host_ver_min, + uint16_t host_ver_cur) +{ + if (!mctp_astlpc_validate_version(bmc_ver_min, bmc_ver_cur, + host_ver_min, host_ver_cur)) + return ASTLPC_VER_BAD; + + if (bmc_ver_cur < host_ver_cur) + return bmc_ver_cur; + + return host_ver_cur; +} + +static int mctp_astlpc_init_host(struct mctp_binding_astlpc *astlpc) +{ + const uint16_t ver_min_be = htobe16(ASTLPC_VER_MIN); + const uint16_t ver_cur_be = htobe16(ASTLPC_VER_CUR); + uint16_t bmc_ver_min, bmc_ver_cur, negotiated; + struct mctp_lpcmap_hdr hdr; + uint8_t status; + int rc; + + rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_STATUS, &status); + if (rc) { + mctp_prwarn("KCS status read failed"); + return rc; + } + + astlpc->kcs_status = status; + + if (!(status & KCS_STATUS_BMC_READY)) + return -EHOSTDOWN; + + mctp_astlpc_lpc_read(astlpc, &hdr, 0, sizeof(hdr)); + + bmc_ver_min = be16toh(hdr.bmc_ver_min); + bmc_ver_cur = be16toh(hdr.bmc_ver_cur); + + /* Calculate the expected value of negotiated_ver */ + negotiated = mctp_astlpc_negotiate_version( + bmc_ver_min, bmc_ver_cur, ASTLPC_VER_MIN, ASTLPC_VER_CUR); + if (!negotiated) { + astlpc_prerr(astlpc, "Cannot negotiate with invalid versions"); + return -EINVAL; + } + + /* Assign protocol ops so we can calculate the packet buffer sizes */ + assert(negotiated < ARRAY_SIZE(astlpc_protocol_version)); + astlpc->proto = &astlpc_protocol_version[negotiated]; + + /* Negotiate packet buffers in v2 style if the BMC supports it */ + if (negotiated >= 2) { + rc = mctp_astlpc_negotiate_layout_host(astlpc); + if (rc < 0) + return rc; + } + + /* Advertise the host's supported protocol versions */ + mctp_astlpc_lpc_write(astlpc, &ver_min_be, + offsetof(struct mctp_lpcmap_hdr, host_ver_min), + sizeof(ver_min_be)); + + mctp_astlpc_lpc_write(astlpc, &ver_cur_be, + offsetof(struct mctp_lpcmap_hdr, host_ver_cur), + sizeof(ver_cur_be)); + + /* Send channel init command */ + rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_DATA, 0x0); + if (rc) { + astlpc_prwarn(astlpc, "KCS write failed"); + } + + /* + * Configure the host so `astlpc->proto->version == 0` holds until we + * receive a subsequent status update from the BMC. Until then, + * `astlpc->proto->version == 0` indicates that we're yet to complete + * the channel initialisation handshake. + * + * When the BMC provides a status update with KCS_STATUS_CHANNEL_ACTIVE + * set we will assign the appropriate protocol ops struct in accordance + * with `negotiated_ver`. + */ + astlpc->proto = &astlpc_protocol_version[ASTLPC_VER_BAD]; + + return rc; +} + +static int mctp_binding_astlpc_start_host(struct mctp_binding *b) +{ + struct mctp_binding_astlpc *astlpc = + container_of(b, struct mctp_binding_astlpc, binding); + + return mctp_astlpc_init_host(astlpc); +} + +static bool __mctp_astlpc_kcs_ready(struct mctp_binding_astlpc *astlpc, + uint8_t status, bool is_write) +{ + bool is_bmc; + bool ready_state; + uint8_t flag; + + is_bmc = (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC); + flag = (is_bmc ^ is_write) ? KCS_STATUS_IBF : KCS_STATUS_OBF; + ready_state = is_write ? 0 : 1; + + return !!(status & flag) == ready_state; +} + +static inline bool +mctp_astlpc_kcs_read_ready(struct mctp_binding_astlpc *astlpc, uint8_t status) +{ + return __mctp_astlpc_kcs_ready(astlpc, status, false); +} + +static inline bool +mctp_astlpc_kcs_write_ready(struct mctp_binding_astlpc *astlpc, uint8_t status) +{ + return __mctp_astlpc_kcs_ready(astlpc, status, true); +} + +static int mctp_astlpc_kcs_send(struct mctp_binding_astlpc *astlpc, + enum mctp_astlpc_cmd data) +{ + uint8_t status; + int rc; + + rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_STATUS, &status); + if (rc) { + astlpc_prwarn(astlpc, "KCS status read failed"); + return -EIO; + } + if (!mctp_astlpc_kcs_write_ready(astlpc, status)) + return -EBUSY; + + rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_DATA, data); + if (rc) { + astlpc_prwarn(astlpc, "KCS data write failed"); + return -EIO; + } + + return 0; +} + +static int mctp_binding_astlpc_tx(struct mctp_binding *b, + struct mctp_pktbuf *pkt) +{ + struct mctp_binding_astlpc *astlpc = binding_to_astlpc(b); + uint32_t len, len_be; + struct mctp_hdr *hdr; + int rc; + + hdr = mctp_pktbuf_hdr(pkt); + len = mctp_pktbuf_size(pkt); + + astlpc_prdebug(astlpc, + "%s: Transmitting %" PRIu32 + "-byte packet (%hhu, %hhu, 0x%hhx)", + __func__, len, hdr->src, hdr->dest, hdr->flags_seq_tag); + + if (len > astlpc->proto->body_size(astlpc->layout.tx.size)) { + astlpc_prwarn(astlpc, "invalid TX len %" PRIu32 ": %" PRIu32, + len, + astlpc->proto->body_size(astlpc->layout.tx.size)); + return -EMSGSIZE; + } + + mctp_binding_set_tx_enabled(b, false); + + len_be = htobe32(len); + mctp_astlpc_lpc_write(astlpc, &len_be, astlpc->layout.tx.offset, + sizeof(len_be)); + + astlpc->proto->pktbuf_protect(pkt); + len = mctp_pktbuf_size(pkt); + + mctp_astlpc_lpc_write(astlpc, hdr, astlpc->layout.tx.offset + 4, len); + + astlpc->layout.tx.state = buffer_state_prepared; + + rc = mctp_astlpc_kcs_send(astlpc, cmd_tx_begin); + if (!rc) + astlpc->layout.tx.state = buffer_state_released; + + return rc == -EBUSY ? 0 : rc; +} + +static uint32_t mctp_astlpc_calculate_mtu(struct mctp_binding_astlpc *astlpc, + struct mctp_astlpc_layout *layout) +{ + uint32_t low, high, limit, rpkt; + + /* Derive the largest MTU the BMC _can_ support */ + low = MIN(astlpc->layout.rx.offset, astlpc->layout.tx.offset); + high = MAX(astlpc->layout.rx.offset, astlpc->layout.tx.offset); + limit = high - low; + + /* Determine the largest MTU the BMC _wants_ to support */ + if (astlpc->requested_mtu) { + uint32_t rmtu = astlpc->requested_mtu; + + rpkt = astlpc->proto->packet_size(MCTP_PACKET_SIZE(rmtu)); + limit = MIN(limit, rpkt); + } + + /* Determine the accepted MTU, applied both directions by convention */ + rpkt = MIN(limit, layout->tx.size); + return MCTP_BODY_SIZE(astlpc->proto->body_size(rpkt)); +} + +static int mctp_astlpc_negotiate_layout_bmc(struct mctp_binding_astlpc *astlpc) +{ + struct mctp_astlpc_layout proposed, pending; + uint32_t sz, mtu; + int rc; + + /* Do we have a valid protocol version? */ + if (!astlpc->proto->version) + return -EINVAL; + + /* Extract the host's proposed layout */ + rc = mctp_astlpc_layout_read(astlpc, &proposed); + if (rc < 0) + return rc; + + /* Do we have a reasonable layout? */ + if (!mctp_astlpc_layout_validate(astlpc, &proposed)) + return -EINVAL; + + /* Negotiate the MTU */ + mtu = mctp_astlpc_calculate_mtu(astlpc, &proposed); + sz = astlpc->proto->packet_size(MCTP_PACKET_SIZE(mtu)); + + /* + * Use symmetric MTUs by convention and to pass constraints in rx/tx + * functions + */ + pending = astlpc->layout; + pending.tx.size = sz; + pending.rx.size = sz; + + if (mctp_astlpc_layout_validate(astlpc, &pending)) { + /* We found a sensible Rx MTU, so honour it */ + astlpc->layout = pending; + + /* Enforce the negotiated MTU */ + rc = mctp_astlpc_layout_write(astlpc, &astlpc->layout); + if (rc < 0) + return rc; + + astlpc_prinfo(astlpc, "Negotiated an MTU of %" PRIu32 " bytes", + mtu); + } else { + astlpc_prwarn(astlpc, "MTU negotiation failed"); + return -EINVAL; + } + + if (astlpc->proto->version >= 2) + astlpc->binding.pkt_size = MCTP_PACKET_SIZE(mtu); + + return 0; +} + +static void mctp_astlpc_init_channel(struct mctp_binding_astlpc *astlpc) +{ + uint16_t negotiated, negotiated_be; + struct mctp_lpcmap_hdr hdr; + uint8_t status; + int rc; + + mctp_astlpc_lpc_read(astlpc, &hdr, 0, sizeof(hdr)); + + /* Version negotiation */ + negotiated = mctp_astlpc_negotiate_version(ASTLPC_VER_MIN, + ASTLPC_VER_CUR, + be16toh(hdr.host_ver_min), + be16toh(hdr.host_ver_cur)); + + /* MTU negotiation requires knowing which protocol we'll use */ + assert(negotiated < ARRAY_SIZE(astlpc_protocol_version)); + astlpc->proto = &astlpc_protocol_version[negotiated]; + + /* Host Rx MTU negotiation: Failure terminates channel init */ + rc = mctp_astlpc_negotiate_layout_bmc(astlpc); + if (rc < 0) + negotiated = ASTLPC_VER_BAD; + + /* Populate the negotiated version */ + negotiated_be = htobe16(negotiated); + mctp_astlpc_lpc_write(astlpc, &negotiated_be, + offsetof(struct mctp_lpcmap_hdr, negotiated_ver), + sizeof(negotiated_be)); + + /* Track buffer ownership */ + astlpc->layout.tx.state = buffer_state_acquired; + astlpc->layout.rx.state = buffer_state_released; + + /* Finalise the configuration */ + status = KCS_STATUS_BMC_READY | KCS_STATUS_OBF; + if (negotiated > 0) { + astlpc_prinfo(astlpc, "Negotiated binding version %" PRIu16, + negotiated); + status |= KCS_STATUS_CHANNEL_ACTIVE; + } else { + astlpc_prerr(astlpc, "Failed to initialise channel"); + } + + mctp_astlpc_kcs_set_status(astlpc, status); + + mctp_binding_set_tx_enabled(&astlpc->binding, + status & KCS_STATUS_CHANNEL_ACTIVE); +} + +static void mctp_astlpc_rx_start(struct mctp_binding_astlpc *astlpc) +{ + struct mctp_pktbuf *pkt; + struct mctp_hdr *hdr; + uint32_t body, packet; + + mctp_astlpc_lpc_read(astlpc, &body, astlpc->layout.rx.offset, + sizeof(body)); + body = be32toh(body); + + if (body > astlpc->proto->body_size(astlpc->layout.rx.size)) { + astlpc_prwarn(astlpc, "invalid RX len 0x%x", body); + return; + } + + if ((size_t)body > astlpc->binding.pkt_size) { + astlpc_prwarn(astlpc, "invalid RX len 0x%x", body); + return; + } + + /* Eliminate the medium-specific header that we just read */ + packet = astlpc->proto->packet_size(body) - 4; + pkt = mctp_pktbuf_alloc(&astlpc->binding, packet); + if (!pkt) { + astlpc_prwarn(astlpc, "unable to allocate pktbuf len 0x%x", + packet); + return; + } + + /* + * Read payload and medium-specific trailer from immediately after the + * medium-specific header. + */ + mctp_astlpc_lpc_read(astlpc, mctp_pktbuf_hdr(pkt), + astlpc->layout.rx.offset + 4, packet); + + astlpc->layout.rx.state = buffer_state_prepared; + + /* Inform the other side of the MCTP interface that we have read + * the packet off the bus before handling the contents of the packet. + */ + if (!mctp_astlpc_kcs_send(astlpc, cmd_rx_complete)) + astlpc->layout.rx.state = buffer_state_released; + + hdr = mctp_pktbuf_hdr(pkt); + if (hdr->ver != 1) { + mctp_pktbuf_free(pkt); + astlpc_prdebug(astlpc, "Dropped packet with invalid version"); + return; + } + + /* + * v3 will validate the CRC32 in the medium-specific trailer and adjust + * the packet size accordingly. On older protocols validation is a no-op + * that always returns true. + */ + if (astlpc->proto->pktbuf_validate(pkt)) { + mctp_bus_rx(&astlpc->binding, pkt); + } else { + /* TODO: Drop any associated assembly */ + mctp_pktbuf_free(pkt); + astlpc_prdebug(astlpc, "Dropped corrupt packet"); + } +} + +static void mctp_astlpc_tx_complete(struct mctp_binding_astlpc *astlpc) +{ + astlpc->layout.tx.state = buffer_state_acquired; + mctp_binding_set_tx_enabled(&astlpc->binding, true); +} + +static int mctp_astlpc_finalise_channel(struct mctp_binding_astlpc *astlpc) +{ + struct mctp_astlpc_layout layout; + uint16_t negotiated; + int rc; + + rc = mctp_astlpc_lpc_read(astlpc, &negotiated, + offsetof(struct mctp_lpcmap_hdr, + negotiated_ver), + sizeof(negotiated)); + if (rc < 0) + return rc; + + negotiated = be16toh(negotiated); + astlpc_prerr(astlpc, "Version negotiation got: %u", negotiated); + + if (negotiated == ASTLPC_VER_BAD || negotiated < ASTLPC_VER_MIN || + negotiated > ASTLPC_VER_CUR) { + astlpc_prerr(astlpc, "Failed to negotiate version, got: %u\n", + negotiated); + return -EINVAL; + } + + assert(negotiated < ARRAY_SIZE(astlpc_protocol_version)); + astlpc->proto = &astlpc_protocol_version[negotiated]; + + rc = mctp_astlpc_layout_read(astlpc, &layout); + if (rc < 0) + return rc; + + if (!mctp_astlpc_layout_validate(astlpc, &layout)) { + mctp_prerr("BMC proposed invalid buffer parameters"); + return -EINVAL; + } + + astlpc->layout = layout; + + if (negotiated >= 2) + astlpc->binding.pkt_size = + astlpc->proto->body_size(astlpc->layout.tx.size); + + /* Track buffer ownership */ + astlpc->layout.tx.state = buffer_state_acquired; + astlpc->layout.rx.state = buffer_state_released; + + return 0; +} + +static int mctp_astlpc_update_channel(struct mctp_binding_astlpc *astlpc, + uint8_t status) +{ + uint8_t updated; + int rc = 0; + + assert(astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST); + + updated = astlpc->kcs_status ^ status; + + astlpc_prdebug(astlpc, "%s: status: 0x%x, update: 0x%x", __func__, + status, updated); + + if (updated & KCS_STATUS_BMC_READY) { + if (status & KCS_STATUS_BMC_READY) { + astlpc->kcs_status = status; + return astlpc->binding.start(&astlpc->binding); + } else { + /* Shut down the channel */ + astlpc->layout.rx.state = buffer_state_idle; + astlpc->layout.tx.state = buffer_state_idle; + mctp_binding_set_tx_enabled(&astlpc->binding, false); + } + } + + if (astlpc->proto->version == 0 || + updated & KCS_STATUS_CHANNEL_ACTIVE) { + bool enable; + + astlpc->layout.rx.state = buffer_state_idle; + astlpc->layout.tx.state = buffer_state_idle; + rc = mctp_astlpc_finalise_channel(astlpc); + enable = (status & KCS_STATUS_CHANNEL_ACTIVE) && rc == 0; + mctp_binding_set_tx_enabled(&astlpc->binding, enable); + } + + astlpc->kcs_status = status; + + return rc; +} + +bool mctp_astlpc_tx_done(struct mctp_binding_astlpc *astlpc) +{ + return astlpc->layout.tx.state == buffer_state_acquired; +} + +int mctp_astlpc_poll(struct mctp_binding_astlpc *astlpc) +{ + uint8_t status, data; + int rc; + + if (astlpc->layout.rx.state == buffer_state_prepared) + if (!mctp_astlpc_kcs_send(astlpc, cmd_rx_complete)) + astlpc->layout.rx.state = buffer_state_released; + + if (astlpc->layout.tx.state == buffer_state_prepared) + if (!mctp_astlpc_kcs_send(astlpc, cmd_tx_begin)) + astlpc->layout.tx.state = buffer_state_released; + + rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_STATUS, &status); + if (rc) { + astlpc_prwarn(astlpc, "KCS read error"); + return -1; + } + + astlpc_prdebug(astlpc, "%s: status: 0x%hhx", __func__, status); + + if (!mctp_astlpc_kcs_read_ready(astlpc, status)) + return 0; + + rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_DATA, &data); + if (rc) { + astlpc_prwarn(astlpc, "KCS data read error"); + return -1; + } + + astlpc_prdebug(astlpc, "%s: data: 0x%hhx", __func__, data); + + if (!astlpc->proto->version && + !(data == cmd_initialise || data == cmd_dummy_value)) { + astlpc_prwarn(astlpc, "Invalid message for binding state: 0x%x", + data); + return 0; + } + + switch (data) { + case cmd_initialise: + mctp_astlpc_init_channel(astlpc); + break; + case cmd_tx_begin: + if (astlpc->layout.rx.state != buffer_state_released) { + astlpc_prerr( + astlpc, + "Protocol error: Invalid Rx buffer state for event %d: %d\n", + data, astlpc->layout.rx.state); + return 0; + } + mctp_astlpc_rx_start(astlpc); + break; + case cmd_rx_complete: + if (astlpc->layout.tx.state != buffer_state_released) { + astlpc_prerr( + astlpc, + "Protocol error: Invalid Tx buffer state for event %d: %d\n", + data, astlpc->layout.tx.state); + return 0; + } + mctp_astlpc_tx_complete(astlpc); + break; + case cmd_dummy_value: + /* No responsibilities for the BMC on 0xff */ + if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST) { + rc = mctp_astlpc_update_channel(astlpc, status); + if (rc < 0) + return rc; + } + break; + default: + astlpc_prwarn(astlpc, "unknown message 0x%x", data); + } + + /* Handle silent loss of bmc-ready */ + if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST) { + if (!(status & KCS_STATUS_BMC_READY && data == cmd_dummy_value)) + return mctp_astlpc_update_channel(astlpc, status); + } + + return rc; +} + +/* allocate and basic initialisation */ +static struct mctp_binding_astlpc *__mctp_astlpc_init(uint8_t mode, + uint32_t mtu) +{ + struct mctp_binding_astlpc *astlpc; + + assert((mode == MCTP_BINDING_ASTLPC_MODE_BMC) || + (mode == MCTP_BINDING_ASTLPC_MODE_HOST)); + + astlpc = __mctp_alloc(sizeof(*astlpc)); + if (!astlpc) + return NULL; + + memset(astlpc, 0, sizeof(*astlpc)); + astlpc->mode = mode; + astlpc->lpc_map = NULL; + astlpc->layout.rx.state = buffer_state_idle; + astlpc->layout.tx.state = buffer_state_idle; + astlpc->requested_mtu = mtu; + astlpc->binding.name = "astlpc"; + astlpc->binding.version = 1; + astlpc->binding.pkt_size = + MCTP_PACKET_SIZE(mtu > MCTP_BTU ? mtu : MCTP_BTU); + astlpc->binding.pkt_header = 4; + astlpc->binding.pkt_trailer = 4; + astlpc->binding.tx = mctp_binding_astlpc_tx; + if (mode == MCTP_BINDING_ASTLPC_MODE_BMC) + astlpc->binding.start = mctp_binding_astlpc_start_bmc; + else if (mode == MCTP_BINDING_ASTLPC_MODE_HOST) + astlpc->binding.start = mctp_binding_astlpc_start_host; + else { + astlpc_prerr(astlpc, "%s: Invalid mode: %d\n", __func__, mode); + __mctp_free(astlpc); + return NULL; + } + + return astlpc; +} + +struct mctp_binding *mctp_binding_astlpc_core(struct mctp_binding_astlpc *b) +{ + return &b->binding; +} + +struct mctp_binding_astlpc * +mctp_astlpc_init(uint8_t mode, uint32_t mtu, void *lpc_map, + const struct mctp_binding_astlpc_ops *ops, void *ops_data) +{ + struct mctp_binding_astlpc *astlpc; + + if (!(mode == MCTP_BINDING_ASTLPC_MODE_BMC || + mode == MCTP_BINDING_ASTLPC_MODE_HOST)) { + mctp_prerr("Unknown binding mode: %u", mode); + return NULL; + } + + astlpc = __mctp_astlpc_init(mode, mtu); + if (!astlpc) + return NULL; + + memcpy(&astlpc->ops, ops, sizeof(astlpc->ops)); + astlpc->ops_data = ops_data; + astlpc->lpc_map = lpc_map; + astlpc->mode = mode; + + return astlpc; +} + +struct mctp_binding_astlpc * +mctp_astlpc_init_ops(const struct mctp_binding_astlpc_ops *ops, void *ops_data, + void *lpc_map) +{ + return mctp_astlpc_init(MCTP_BINDING_ASTLPC_MODE_BMC, MCTP_BTU, lpc_map, + ops, ops_data); +} + +void mctp_astlpc_destroy(struct mctp_binding_astlpc *astlpc) +{ + /* Clear channel-active and bmc-ready */ + if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC) + mctp_astlpc_kcs_set_status(astlpc, 0); + __mctp_free(astlpc); +} + +#ifdef MCTP_HAVE_FILEIO + +static int mctp_astlpc_init_fileio_lpc(struct mctp_binding_astlpc *astlpc) +{ + struct aspeed_lpc_ctrl_mapping map = { + .window_type = ASPEED_LPC_CTRL_WINDOW_MEMORY, + .window_id = 0, /* There's only one */ + .flags = 0, + .addr = 0, + .offset = 0, + .size = 0 + }; + void *lpc_map_base; + int fd, rc; + + fd = open(lpc_path, O_RDWR | O_SYNC); + if (fd < 0) { + astlpc_prwarn(astlpc, "LPC open (%s) failed", lpc_path); + return -1; + } + + rc = ioctl(fd, ASPEED_LPC_CTRL_IOCTL_GET_SIZE, &map); + if (rc) { + astlpc_prwarn(astlpc, "LPC GET_SIZE failed"); + close(fd); + return -1; + } + + /* + * 🚨🚨🚨 + * + * Decouple ourselves from hiomapd[1] (another user of the FW2AHB) by + * mapping the FW2AHB to the reserved memory here as well. + * + * It's not possible to use the MCTP ASTLPC binding on machines that + * need the FW2AHB bridge mapped anywhere except to the reserved memory + * (e.g. the host SPI NOR). + * + * [1] https://github.com/openbmc/hiomapd/ + * + * 🚨🚨🚨 + * + * The following calculation must align with what's going on in + * hiomapd's lpc.c so as not to disrupt its behaviour: + * + * https://github.com/openbmc/hiomapd/blob/5ff50e3cbd7702aefc185264e4adfb9952040575/lpc.c#L68 + * + * 🚨🚨🚨 + */ + + /* Map the reserved memory at the top of the 28-bit LPC firmware address space */ + map.addr = 0x0FFFFFFF & -map.size; + astlpc_prinfo( + astlpc, + "Configuring FW2AHB to map reserved memory at 0x%08x for 0x%x in the LPC FW cycle address-space", + map.addr, map.size); + + rc = ioctl(fd, ASPEED_LPC_CTRL_IOCTL_MAP, &map); + if (rc) { + astlpc_prwarn(astlpc, + "Failed to map FW2AHB to reserved memory"); + close(fd); + return -1; + } + + /* Map the reserved memory into our address space */ + lpc_map_base = + mmap(NULL, map.size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0); + if (lpc_map_base == MAP_FAILED) { + astlpc_prwarn(astlpc, "LPC mmap failed"); + rc = -1; + } else { + astlpc->lpc_map = lpc_map_base + map.size - LPC_WIN_SIZE; + } + + close(fd); + + return rc; +} + +static int mctp_astlpc_init_fileio_kcs(struct mctp_binding_astlpc *astlpc) +{ + astlpc->kcs_fd = open(kcs_path, O_RDWR); + if (astlpc->kcs_fd < 0) + return -1; + + return 0; +} + +static int __mctp_astlpc_fileio_kcs_read(void *arg, + enum mctp_binding_astlpc_kcs_reg reg, + uint8_t *val) +{ + struct mctp_binding_astlpc *astlpc = arg; + off_t offset = reg; + int rc; + + rc = pread(astlpc->kcs_fd, val, 1, offset); + + return rc == 1 ? 0 : -1; +} + +static int __mctp_astlpc_fileio_kcs_write(void *arg, + enum mctp_binding_astlpc_kcs_reg reg, + uint8_t val) +{ + struct mctp_binding_astlpc *astlpc = arg; + off_t offset = reg; + int rc; + + rc = pwrite(astlpc->kcs_fd, &val, 1, offset); + + return rc == 1 ? 0 : -1; +} + +int mctp_astlpc_init_pollfd(struct mctp_binding_astlpc *astlpc, + struct pollfd *pollfd) +{ + bool release; + + pollfd->fd = astlpc->kcs_fd; + pollfd->events = 0; + + release = astlpc->layout.rx.state == buffer_state_prepared || + astlpc->layout.tx.state == buffer_state_prepared; + + pollfd->events = release ? POLLOUT : POLLIN; + + return 0; +} + +struct mctp_binding_astlpc *mctp_astlpc_init_fileio(void) +{ + struct mctp_binding_astlpc *astlpc; + int rc; + + /* + * If we're doing file IO then we're very likely not running + * freestanding, so lets assume that we're on the BMC side. + * + * Requesting an MTU of 0 requests the largest possible MTU, whatever + * value that might take. + */ + astlpc = __mctp_astlpc_init(MCTP_BINDING_ASTLPC_MODE_BMC, 0); + if (!astlpc) + return NULL; + + /* Set internal operations for kcs. We use direct accesses to the lpc + * map area */ + astlpc->ops.kcs_read = __mctp_astlpc_fileio_kcs_read; + astlpc->ops.kcs_write = __mctp_astlpc_fileio_kcs_write; + astlpc->ops_data = astlpc; + + rc = mctp_astlpc_init_fileio_lpc(astlpc); + if (rc) { + free(astlpc); + return NULL; + } + + rc = mctp_astlpc_init_fileio_kcs(astlpc); + if (rc) { + free(astlpc); + return NULL; + } + + return astlpc; +} +#else +struct mctp_binding_astlpc *mctp_astlpc_init_fileio(void) +{ + mctp_prlog(MCTP_LOG_ERR, "%s: Missing support for file IO", __func__); + return NULL; +} + +int mctp_astlpc_init_pollfd(struct mctp_binding_astlpc *astlpc __unused, + struct pollfd *pollfd __unused) +{ + mctp_prlog(MCTP_LOG_ERR, "%s: Missing support for file IO", __func__); + return -1; +} +#endif |