openocd/src/jtag/drivers/xlnx-pcie-xvc.c

707 lines
17 KiB
C

/* SPDX-License-Identifier: GPL-2.0-only */
/*
* Copyright (c) 2019 Google, LLC.
* Author: Moritz Fischer <moritzf@google.com>
*/
#ifdef HAVE_CONFIG_H
#include "config.h"
#endif
#include <stdint.h>
#include <stdlib.h>
#include <math.h>
#include <unistd.h>
#include <linux/pci.h>
#include <jtag/interface.h>
#include <jtag/swd.h>
#include <jtag/commands.h>
#include <helper/replacements.h>
#include <helper/bits.h>
/* Available only from kernel v4.10 */
#ifndef PCI_CFG_SPACE_EXP_SIZE
#define PCI_CFG_SPACE_EXP_SIZE 4096
#endif
#define PCIE_EXT_CAP_LST 0x100
#define XLNX_XVC_EXT_CAP 0x00
#define XLNX_XVC_VSEC_HDR 0x04
#define XLNX_XVC_LEN_REG 0x0C
#define XLNX_XVC_TMS_REG 0x10
#define XLNX_XVC_TDX_REG 0x14
#define XLNX_XVC_CAP_SIZE 0x20
#define XLNX_XVC_VSEC_ID 0x8
#define XLNX_XVC_MAX_BITS 0x20
#define MASK_ACK(x) (((x) >> 9) & 0x7)
#define MASK_PAR(x) ((int)((x) & 0x1))
struct xlnx_pcie_xvc {
int fd;
unsigned offset;
char *device;
};
static struct xlnx_pcie_xvc xlnx_pcie_xvc_state;
static struct xlnx_pcie_xvc *xlnx_pcie_xvc = &xlnx_pcie_xvc_state;
static int xlnx_pcie_xvc_read_reg(const int offset, uint32_t *val)
{
uint32_t res;
int err;
/* Note: This should be ok endianness-wise because by going
* through sysfs the kernel does the conversion in the config
* space accessor functions
*/
err = pread(xlnx_pcie_xvc->fd, &res, sizeof(res),
xlnx_pcie_xvc->offset + offset);
if (err != sizeof(res)) {
LOG_ERROR("Failed to read offset %x", offset);
return ERROR_JTAG_DEVICE_ERROR;
}
if (val)
*val = res;
return ERROR_OK;
}
static int xlnx_pcie_xvc_write_reg(const int offset, const uint32_t val)
{
int err;
/* Note: This should be ok endianness-wise because by going
* through sysfs the kernel does the conversion in the config
* space accessor functions
*/
err = pwrite(xlnx_pcie_xvc->fd, &val, sizeof(val),
xlnx_pcie_xvc->offset + offset);
if (err != sizeof(val)) {
LOG_ERROR("Failed to write offset: %x with value: %" PRIx32,
offset, val);
return ERROR_JTAG_DEVICE_ERROR;
}
return ERROR_OK;
}
static int xlnx_pcie_xvc_transact(size_t num_bits, uint32_t tms, uint32_t tdi,
uint32_t *tdo)
{
int err;
err = xlnx_pcie_xvc_write_reg(XLNX_XVC_LEN_REG, num_bits);
if (err != ERROR_OK)
return err;
err = xlnx_pcie_xvc_write_reg(XLNX_XVC_TMS_REG, tms);
if (err != ERROR_OK)
return err;
err = xlnx_pcie_xvc_write_reg(XLNX_XVC_TDX_REG, tdi);
if (err != ERROR_OK)
return err;
err = xlnx_pcie_xvc_read_reg(XLNX_XVC_TDX_REG, tdo);
if (err != ERROR_OK)
return err;
if (tdo)
LOG_DEBUG_IO("Transact num_bits: %zu, tms: %" PRIx32 ", tdi: %" PRIx32 ", tdo: %" PRIx32,
num_bits, tms, tdi, *tdo);
else
LOG_DEBUG_IO("Transact num_bits: %zu, tms: %" PRIx32 ", tdi: %" PRIx32 ", tdo: <null>",
num_bits, tms, tdi);
return ERROR_OK;
}
static int xlnx_pcie_xvc_execute_stableclocks(struct jtag_command *cmd)
{
int tms = tap_get_state() == TAP_RESET ? 1 : 0;
size_t left = cmd->cmd.stableclocks->num_cycles;
size_t write;
int err;
LOG_DEBUG("stableclocks %i cycles", cmd->cmd.runtest->num_cycles);
while (left) {
write = MIN(XLNX_XVC_MAX_BITS, left);
err = xlnx_pcie_xvc_transact(write, tms, 0, NULL);
if (err != ERROR_OK)
return err;
left -= write;
};
return ERROR_OK;
}
static int xlnx_pcie_xvc_execute_statemove(size_t skip)
{
uint8_t tms_scan = tap_get_tms_path(tap_get_state(),
tap_get_end_state());
int tms_count = tap_get_tms_path_len(tap_get_state(),
tap_get_end_state());
int err;
LOG_DEBUG("statemove starting at (skip: %zu) %s end in %s", skip,
tap_state_name(tap_get_state()),
tap_state_name(tap_get_end_state()));
err = xlnx_pcie_xvc_transact(tms_count - skip, tms_scan >> skip, 0, NULL);
if (err != ERROR_OK)
return err;
tap_set_state(tap_get_end_state());
return ERROR_OK;
}
static int xlnx_pcie_xvc_execute_runtest(struct jtag_command *cmd)
{
int err = ERROR_OK;
LOG_DEBUG("runtest %i cycles, end in %i",
cmd->cmd.runtest->num_cycles,
cmd->cmd.runtest->end_state);
tap_state_t tmp_state = tap_get_end_state();
if (tap_get_state() != TAP_IDLE) {
tap_set_end_state(TAP_IDLE);
err = xlnx_pcie_xvc_execute_statemove(0);
if (err != ERROR_OK)
return err;
};
size_t left = cmd->cmd.runtest->num_cycles;
size_t write;
while (left) {
write = MIN(XLNX_XVC_MAX_BITS, left);
err = xlnx_pcie_xvc_transact(write, 0, 0, NULL);
if (err != ERROR_OK)
return err;
left -= write;
};
tap_set_end_state(tmp_state);
if (tap_get_state() != tap_get_end_state())
err = xlnx_pcie_xvc_execute_statemove(0);
return err;
}
static int xlnx_pcie_xvc_execute_pathmove(struct jtag_command *cmd)
{
size_t num_states = cmd->cmd.pathmove->num_states;
tap_state_t *path = cmd->cmd.pathmove->path;
int err = ERROR_OK;
size_t i;
LOG_DEBUG("pathmove: %i states, end in %i",
cmd->cmd.pathmove->num_states,
cmd->cmd.pathmove->path[cmd->cmd.pathmove->num_states - 1]);
for (i = 0; i < num_states; i++) {
if (path[i] == tap_state_transition(tap_get_state(), false)) {
err = xlnx_pcie_xvc_transact(1, 1, 0, NULL);
} else if (path[i] == tap_state_transition(tap_get_state(), true)) {
err = xlnx_pcie_xvc_transact(1, 0, 0, NULL);
} else {
LOG_ERROR("BUG: %s -> %s isn't a valid TAP transition.",
tap_state_name(tap_get_state()),
tap_state_name(path[i]));
err = ERROR_JTAG_QUEUE_FAILED;
}
if (err != ERROR_OK)
return err;
tap_set_state(path[i]);
}
tap_set_end_state(tap_get_state());
return ERROR_OK;
}
static int xlnx_pcie_xvc_execute_scan(struct jtag_command *cmd)
{
enum scan_type type = jtag_scan_type(cmd->cmd.scan);
tap_state_t saved_end_state = cmd->cmd.scan->end_state;
bool ir_scan = cmd->cmd.scan->ir_scan;
uint32_t tdi, tms, tdo;
uint8_t *buf, *rd_ptr;
int err, scan_size;
size_t write;
size_t left;
scan_size = jtag_build_buffer(cmd->cmd.scan, &buf);
rd_ptr = buf;
LOG_DEBUG("%s scan type %d %d bits; starts in %s end in %s",
(cmd->cmd.scan->ir_scan) ? "IR" : "DR", type, scan_size,
tap_state_name(tap_get_state()),
tap_state_name(cmd->cmd.scan->end_state));
/* If we're in TAP_DR_SHIFT state but need to do a IR_SCAN or
* vice-versa, do a statemove to corresponding other state, then restore
* end state
*/
if (ir_scan && tap_get_state() != TAP_IRSHIFT) {
tap_set_end_state(TAP_IRSHIFT);
err = xlnx_pcie_xvc_execute_statemove(0);
if (err != ERROR_OK)
goto out_err;
tap_set_end_state(saved_end_state);
} else if (!ir_scan && (tap_get_state() != TAP_DRSHIFT)) {
tap_set_end_state(TAP_DRSHIFT);
err = xlnx_pcie_xvc_execute_statemove(0);
if (err != ERROR_OK)
goto out_err;
tap_set_end_state(saved_end_state);
}
left = scan_size;
while (left) {
write = MIN(XLNX_XVC_MAX_BITS, left);
/* the last TMS should be a 1, to leave the state */
tms = left <= XLNX_XVC_MAX_BITS ? BIT(write - 1) : 0;
tdi = (type != SCAN_IN) ? buf_get_u32(rd_ptr, 0, write) : 0;
err = xlnx_pcie_xvc_transact(write, tms, tdi, type != SCAN_OUT ?
&tdo : NULL);
if (err != ERROR_OK)
goto out_err;
left -= write;
if (type != SCAN_OUT)
buf_set_u32(rd_ptr, 0, write, tdo);
rd_ptr += sizeof(uint32_t);
};
err = jtag_read_buffer(buf, cmd->cmd.scan);
free(buf);
if (tap_get_state() != tap_get_end_state())
err = xlnx_pcie_xvc_execute_statemove(1);
return err;
out_err:
free(buf);
return err;
}
static void xlnx_pcie_xvc_execute_reset(struct jtag_command *cmd)
{
LOG_DEBUG("reset trst: %i srst: %i", cmd->cmd.reset->trst,
cmd->cmd.reset->srst);
}
static void xlnx_pcie_xvc_execute_sleep(struct jtag_command *cmd)
{
LOG_DEBUG("sleep %" PRIu32 "", cmd->cmd.sleep->us);
usleep(cmd->cmd.sleep->us);
}
static int xlnx_pcie_xvc_execute_tms(struct jtag_command *cmd)
{
const size_t num_bits = cmd->cmd.tms->num_bits;
const uint8_t *bits = cmd->cmd.tms->bits;
size_t left, write;
uint32_t tms;
int err;
LOG_DEBUG("execute tms %zu", num_bits);
left = num_bits;
while (left) {
write = MIN(XLNX_XVC_MAX_BITS, left);
tms = buf_get_u32(bits, 0, write);
err = xlnx_pcie_xvc_transact(write, tms, 0, NULL);
if (err != ERROR_OK)
return err;
left -= write;
bits += 4;
};
return ERROR_OK;
}
static int xlnx_pcie_xvc_execute_command(struct jtag_command *cmd)
{
LOG_DEBUG("%s: cmd->type: %u", __func__, cmd->type);
switch (cmd->type) {
case JTAG_STABLECLOCKS:
return xlnx_pcie_xvc_execute_stableclocks(cmd);
case JTAG_RUNTEST:
return xlnx_pcie_xvc_execute_runtest(cmd);
case JTAG_TLR_RESET:
tap_set_end_state(cmd->cmd.statemove->end_state);
return xlnx_pcie_xvc_execute_statemove(0);
case JTAG_PATHMOVE:
return xlnx_pcie_xvc_execute_pathmove(cmd);
case JTAG_SCAN:
return xlnx_pcie_xvc_execute_scan(cmd);
case JTAG_RESET:
xlnx_pcie_xvc_execute_reset(cmd);
break;
case JTAG_SLEEP:
xlnx_pcie_xvc_execute_sleep(cmd);
break;
case JTAG_TMS:
return xlnx_pcie_xvc_execute_tms(cmd);
default:
LOG_ERROR("BUG: Unknown JTAG command type encountered.");
return ERROR_JTAG_QUEUE_FAILED;
}
return ERROR_OK;
}
static int xlnx_pcie_xvc_execute_queue(void)
{
struct jtag_command *cmd = jtag_command_queue;
int ret;
while (cmd) {
ret = xlnx_pcie_xvc_execute_command(cmd);
if (ret != ERROR_OK)
return ret;
cmd = cmd->next;
}
return ERROR_OK;
}
static int xlnx_pcie_xvc_init(void)
{
char filename[PATH_MAX];
uint32_t cap, vh;
int err;
snprintf(filename, PATH_MAX, "/sys/bus/pci/devices/%s/config",
xlnx_pcie_xvc->device);
xlnx_pcie_xvc->fd = open(filename, O_RDWR | O_SYNC);
if (xlnx_pcie_xvc->fd < 0) {
LOG_ERROR("Failed to open device: %s", filename);
return ERROR_JTAG_INIT_FAILED;
}
LOG_INFO("Scanning PCIe device %s's for Xilinx XVC/PCIe ...",
xlnx_pcie_xvc->device);
/* Parse the PCIe extended capability list and try to find
* vendor specific header */
xlnx_pcie_xvc->offset = PCIE_EXT_CAP_LST;
while (xlnx_pcie_xvc->offset <= PCI_CFG_SPACE_EXP_SIZE - sizeof(cap) &&
xlnx_pcie_xvc->offset >= PCIE_EXT_CAP_LST) {
err = xlnx_pcie_xvc_read_reg(XLNX_XVC_EXT_CAP, &cap);
if (err != ERROR_OK)
return err;
LOG_DEBUG("Checking capability at 0x%x; id=0x%04" PRIx32 " version=0x%" PRIx32 " next=0x%" PRIx32,
xlnx_pcie_xvc->offset,
PCI_EXT_CAP_ID(cap),
PCI_EXT_CAP_VER(cap),
PCI_EXT_CAP_NEXT(cap));
if (PCI_EXT_CAP_ID(cap) == PCI_EXT_CAP_ID_VNDR) {
err = xlnx_pcie_xvc_read_reg(XLNX_XVC_VSEC_HDR, &vh);
if (err != ERROR_OK)
return err;
LOG_DEBUG("Checking possible match at 0x%x; id: 0x%" PRIx32 "; rev: 0x%" PRIx32 "; length: 0x%" PRIx32,
xlnx_pcie_xvc->offset,
PCI_VNDR_HEADER_ID(vh),
PCI_VNDR_HEADER_REV(vh),
PCI_VNDR_HEADER_LEN(vh));
if ((PCI_VNDR_HEADER_ID(vh) == XLNX_XVC_VSEC_ID) &&
(PCI_VNDR_HEADER_LEN(vh) == XLNX_XVC_CAP_SIZE))
break;
}
xlnx_pcie_xvc->offset = PCI_EXT_CAP_NEXT(cap);
}
if ((xlnx_pcie_xvc->offset > PCI_CFG_SPACE_EXP_SIZE - XLNX_XVC_CAP_SIZE) ||
xlnx_pcie_xvc->offset < PCIE_EXT_CAP_LST) {
close(xlnx_pcie_xvc->fd);
return ERROR_JTAG_INIT_FAILED;
}
LOG_INFO("Found Xilinx XVC/PCIe capability at offset: 0x%x", xlnx_pcie_xvc->offset);
return ERROR_OK;
}
static int xlnx_pcie_xvc_quit(void)
{
int err;
err = close(xlnx_pcie_xvc->fd);
if (err)
return err;
return ERROR_OK;
}
COMMAND_HANDLER(xlnx_pcie_xvc_handle_config_command)
{
if (CMD_ARGC < 1)
return ERROR_COMMAND_SYNTAX_ERROR;
/* we can't really free this in a safe manner, so at least
* limit the memory we're leaking by freeing the old one first
* before allocating a new one ...
*/
free(xlnx_pcie_xvc->device);
xlnx_pcie_xvc->device = strdup(CMD_ARGV[0]);
return ERROR_OK;
}
static const struct command_registration xlnx_pcie_xvc_subcommand_handlers[] = {
{
.name = "config",
.handler = xlnx_pcie_xvc_handle_config_command,
.mode = COMMAND_CONFIG,
.help = "Configure XVC/PCIe JTAG adapter",
.usage = "device",
},
COMMAND_REGISTRATION_DONE
};
static const struct command_registration xlnx_pcie_xvc_command_handlers[] = {
{
.name = "xlnx_pcie_xvc",
.mode = COMMAND_ANY,
.help = "perform xlnx_pcie_xvc management",
.chain = xlnx_pcie_xvc_subcommand_handlers,
.usage = "",
},
COMMAND_REGISTRATION_DONE
};
static struct jtag_interface xlnx_pcie_xvc_jtag_ops = {
.execute_queue = &xlnx_pcie_xvc_execute_queue,
};
static int xlnx_pcie_xvc_swd_sequence(const uint8_t *seq, size_t length)
{
size_t left, write;
uint32_t send;
int err;
left = length;
while (left) {
write = MIN(XLNX_XVC_MAX_BITS, left);
send = buf_get_u32(seq, 0, write);
err = xlnx_pcie_xvc_transact(write, send, 0, NULL);
if (err != ERROR_OK)
return err;
left -= write;
seq += sizeof(uint32_t);
};
return ERROR_OK;
}
static int xlnx_pcie_xvc_swd_switch_seq(enum swd_special_seq seq)
{
switch (seq) {
case LINE_RESET:
LOG_DEBUG("SWD line reset");
return xlnx_pcie_xvc_swd_sequence(swd_seq_line_reset,
swd_seq_line_reset_len);
case JTAG_TO_SWD:
LOG_DEBUG("JTAG-to-SWD");
return xlnx_pcie_xvc_swd_sequence(swd_seq_jtag_to_swd,
swd_seq_jtag_to_swd_len);
case SWD_TO_JTAG:
LOG_DEBUG("SWD-to-JTAG");
return xlnx_pcie_xvc_swd_sequence(swd_seq_swd_to_jtag,
swd_seq_swd_to_jtag_len);
default:
LOG_ERROR("Sequence %d not supported", seq);
return ERROR_FAIL;
}
return ERROR_OK;
}
static int queued_retval;
static void xlnx_pcie_xvc_swd_write_reg(uint8_t cmd, uint32_t value,
uint32_t ap_delay_clk);
static void swd_clear_sticky_errors(void)
{
xlnx_pcie_xvc_swd_write_reg(swd_cmd(false, false, DP_ABORT),
STKCMPCLR | STKERRCLR | WDERRCLR | ORUNERRCLR, 0);
}
static void xlnx_pcie_xvc_swd_read_reg(uint8_t cmd, uint32_t *value,
uint32_t ap_delay_clk)
{
uint32_t res, ack, rpar;
int err;
assert(cmd & SWD_CMD_RNW);
cmd |= SWD_CMD_START | SWD_CMD_PARK;
/* cmd + ack */
err = xlnx_pcie_xvc_transact(12, cmd, 0, &res);
if (err != ERROR_OK)
goto err_out;
ack = MASK_ACK(res);
/* read data */
err = xlnx_pcie_xvc_transact(32, 0, 0, &res);
if (err != ERROR_OK)
goto err_out;
/* parity + trn */
err = xlnx_pcie_xvc_transact(2, 0, 0, &rpar);
if (err != ERROR_OK)
goto err_out;
LOG_DEBUG("%s %s %s reg %X = %08"PRIx32,
ack == SWD_ACK_OK ? "OK" : ack == SWD_ACK_WAIT ?
"WAIT" : ack == SWD_ACK_FAULT ? "FAULT" : "JUNK",
cmd & SWD_CMD_APNDP ? "AP" : "DP",
cmd & SWD_CMD_RNW ? "read" : "write",
(cmd & SWD_CMD_A32) >> 1,
res);
switch (ack) {
case SWD_ACK_OK:
if (MASK_PAR(rpar) != parity_u32(res)) {
LOG_DEBUG_IO("Wrong parity detected");
queued_retval = ERROR_FAIL;
return;
}
if (value)
*value = res;
if (cmd & SWD_CMD_APNDP)
err = xlnx_pcie_xvc_transact(ap_delay_clk, 0, 0, NULL);
queued_retval = err;
return;
case SWD_ACK_WAIT:
LOG_DEBUG_IO("SWD_ACK_WAIT");
swd_clear_sticky_errors();
return;
case SWD_ACK_FAULT:
LOG_DEBUG_IO("SWD_ACK_FAULT");
queued_retval = ack;
return;
default:
LOG_DEBUG_IO("No valid acknowledge: ack=%02"PRIx32, ack);
queued_retval = ack;
return;
}
err_out:
queued_retval = err;
}
static void xlnx_pcie_xvc_swd_write_reg(uint8_t cmd, uint32_t value,
uint32_t ap_delay_clk)
{
uint32_t res, ack;
int err;
assert(!(cmd & SWD_CMD_RNW));
cmd |= SWD_CMD_START | SWD_CMD_PARK;
/* cmd + trn + ack */
err = xlnx_pcie_xvc_transact(13, cmd, 0, &res);
if (err != ERROR_OK)
goto err_out;
ack = MASK_ACK(res);
/* write data */
err = xlnx_pcie_xvc_transact(32, value, 0, NULL);
if (err != ERROR_OK)
goto err_out;
/* parity + trn */
err = xlnx_pcie_xvc_transact(2, parity_u32(value), 0, NULL);
if (err != ERROR_OK)
goto err_out;
LOG_DEBUG("%s %s %s reg %X = %08"PRIx32,
ack == SWD_ACK_OK ? "OK" : ack == SWD_ACK_WAIT ?
"WAIT" : ack == SWD_ACK_FAULT ? "FAULT" : "JUNK",
cmd & SWD_CMD_APNDP ? "AP" : "DP",
cmd & SWD_CMD_RNW ? "read" : "write",
(cmd & SWD_CMD_A32) >> 1,
value);
switch (ack) {
case SWD_ACK_OK:
if (cmd & SWD_CMD_APNDP)
err = xlnx_pcie_xvc_transact(ap_delay_clk, 0, 0, NULL);
queued_retval = err;
return;
case SWD_ACK_WAIT:
LOG_DEBUG_IO("SWD_ACK_WAIT");
swd_clear_sticky_errors();
return;
case SWD_ACK_FAULT:
LOG_DEBUG_IO("SWD_ACK_FAULT");
queued_retval = ack;
return;
default:
LOG_DEBUG_IO("No valid acknowledge: ack=%02"PRIx32, ack);
queued_retval = ack;
return;
}
err_out:
queued_retval = err;
}
static int xlnx_pcie_xvc_swd_run_queue(void)
{
int err;
/* we want at least 8 idle cycles between each transaction */
err = xlnx_pcie_xvc_transact(8, 0, 0, NULL);
if (err != ERROR_OK)
return err;
err = queued_retval;
queued_retval = ERROR_OK;
LOG_DEBUG("SWD queue return value: %02x", err);
return err;
}
static int xlnx_pcie_xvc_swd_init(void)
{
return ERROR_OK;
}
static const struct swd_driver xlnx_pcie_xvc_swd_ops = {
.init = xlnx_pcie_xvc_swd_init,
.switch_seq = xlnx_pcie_xvc_swd_switch_seq,
.read_reg = xlnx_pcie_xvc_swd_read_reg,
.write_reg = xlnx_pcie_xvc_swd_write_reg,
.run = xlnx_pcie_xvc_swd_run_queue,
};
static const char * const xlnx_pcie_xvc_transports[] = { "jtag", "swd", NULL };
struct adapter_driver xlnx_pcie_xvc_adapter_driver = {
.name = "xlnx_pcie_xvc",
.transports = xlnx_pcie_xvc_transports,
.commands = xlnx_pcie_xvc_command_handlers,
.init = &xlnx_pcie_xvc_init,
.quit = &xlnx_pcie_xvc_quit,
.jtag_ops = &xlnx_pcie_xvc_jtag_ops,
.swd_ops = &xlnx_pcie_xvc_swd_ops,
};