384 lines
12 KiB
C
384 lines
12 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (c) 2018-2020, The Linux Foundation. All rights reserved.
|
|
*
|
|
*/
|
|
|
|
#ifndef _MHI_INT_H
|
|
#define _MHI_INT_H
|
|
|
|
#include "../common.h"
|
|
|
|
extern struct bus_type mhi_bus_type;
|
|
|
|
/* Host request register */
|
|
#define MHI_SOC_RESET_REQ_OFFSET 0xb0
|
|
#define MHI_SOC_RESET_REQ BIT(0)
|
|
|
|
#define SOC_HW_VERSION_OFFS 0x224
|
|
#define SOC_HW_VERSION_FAM_NUM_BMSK GENMASK(31, 28)
|
|
#define SOC_HW_VERSION_DEV_NUM_BMSK GENMASK(27, 16)
|
|
#define SOC_HW_VERSION_MAJOR_VER_BMSK GENMASK(15, 8)
|
|
#define SOC_HW_VERSION_MINOR_VER_BMSK GENMASK(7, 0)
|
|
|
|
struct mhi_ctxt {
|
|
struct mhi_event_ctxt *er_ctxt;
|
|
struct mhi_chan_ctxt *chan_ctxt;
|
|
struct mhi_cmd_ctxt *cmd_ctxt;
|
|
dma_addr_t er_ctxt_addr;
|
|
dma_addr_t chan_ctxt_addr;
|
|
dma_addr_t cmd_ctxt_addr;
|
|
};
|
|
|
|
struct bhi_vec_entry {
|
|
u64 dma_addr;
|
|
u64 size;
|
|
};
|
|
|
|
enum mhi_ch_state_type {
|
|
MHI_CH_STATE_TYPE_RESET,
|
|
MHI_CH_STATE_TYPE_STOP,
|
|
MHI_CH_STATE_TYPE_START,
|
|
MHI_CH_STATE_TYPE_MAX,
|
|
};
|
|
|
|
extern const char * const mhi_ch_state_type_str[MHI_CH_STATE_TYPE_MAX];
|
|
#define TO_CH_STATE_TYPE_STR(state) (((state) >= MHI_CH_STATE_TYPE_MAX) ? \
|
|
"INVALID_STATE" : \
|
|
mhi_ch_state_type_str[(state)])
|
|
|
|
#define MHI_INVALID_BRSTMODE(mode) (mode != MHI_DB_BRST_DISABLE && \
|
|
mode != MHI_DB_BRST_ENABLE)
|
|
|
|
extern const char * const mhi_ee_str[MHI_EE_MAX];
|
|
#define TO_MHI_EXEC_STR(ee) (((ee) >= MHI_EE_MAX) ? \
|
|
"INVALID_EE" : mhi_ee_str[ee])
|
|
|
|
#define MHI_IN_PBL(ee) (ee == MHI_EE_PBL || ee == MHI_EE_PTHRU || \
|
|
ee == MHI_EE_EDL)
|
|
#define MHI_POWER_UP_CAPABLE(ee) (MHI_IN_PBL(ee) || ee == MHI_EE_AMSS)
|
|
#define MHI_FW_LOAD_CAPABLE(ee) (ee == MHI_EE_PBL || ee == MHI_EE_EDL)
|
|
#define MHI_IN_MISSION_MODE(ee) (ee == MHI_EE_AMSS || ee == MHI_EE_WFW || \
|
|
ee == MHI_EE_FP)
|
|
|
|
enum dev_st_transition {
|
|
DEV_ST_TRANSITION_PBL,
|
|
DEV_ST_TRANSITION_READY,
|
|
DEV_ST_TRANSITION_SBL,
|
|
DEV_ST_TRANSITION_MISSION_MODE,
|
|
DEV_ST_TRANSITION_FP,
|
|
DEV_ST_TRANSITION_SYS_ERR,
|
|
DEV_ST_TRANSITION_DISABLE,
|
|
DEV_ST_TRANSITION_MAX,
|
|
};
|
|
|
|
extern const char * const dev_state_tran_str[DEV_ST_TRANSITION_MAX];
|
|
#define TO_DEV_STATE_TRANS_STR(state) (((state) >= DEV_ST_TRANSITION_MAX) ? \
|
|
"INVALID_STATE" : dev_state_tran_str[state])
|
|
|
|
/* internal power states */
|
|
enum mhi_pm_state {
|
|
MHI_PM_STATE_DISABLE,
|
|
MHI_PM_STATE_POR,
|
|
MHI_PM_STATE_M0,
|
|
MHI_PM_STATE_M2,
|
|
MHI_PM_STATE_M3_ENTER,
|
|
MHI_PM_STATE_M3,
|
|
MHI_PM_STATE_M3_EXIT,
|
|
MHI_PM_STATE_FW_DL_ERR,
|
|
MHI_PM_STATE_SYS_ERR_DETECT,
|
|
MHI_PM_STATE_SYS_ERR_PROCESS,
|
|
MHI_PM_STATE_SHUTDOWN_PROCESS,
|
|
MHI_PM_STATE_LD_ERR_FATAL_DETECT,
|
|
MHI_PM_STATE_MAX
|
|
};
|
|
|
|
#define MHI_PM_DISABLE BIT(0)
|
|
#define MHI_PM_POR BIT(1)
|
|
#define MHI_PM_M0 BIT(2)
|
|
#define MHI_PM_M2 BIT(3)
|
|
#define MHI_PM_M3_ENTER BIT(4)
|
|
#define MHI_PM_M3 BIT(5)
|
|
#define MHI_PM_M3_EXIT BIT(6)
|
|
/* firmware download failure state */
|
|
#define MHI_PM_FW_DL_ERR BIT(7)
|
|
#define MHI_PM_SYS_ERR_DETECT BIT(8)
|
|
#define MHI_PM_SYS_ERR_PROCESS BIT(9)
|
|
#define MHI_PM_SHUTDOWN_PROCESS BIT(10)
|
|
/* link not accessible */
|
|
#define MHI_PM_LD_ERR_FATAL_DETECT BIT(11)
|
|
|
|
#define MHI_REG_ACCESS_VALID(pm_state) ((pm_state & (MHI_PM_POR | MHI_PM_M0 | \
|
|
MHI_PM_M2 | MHI_PM_M3_ENTER | MHI_PM_M3_EXIT | \
|
|
MHI_PM_SYS_ERR_DETECT | MHI_PM_SYS_ERR_PROCESS | \
|
|
MHI_PM_SHUTDOWN_PROCESS | MHI_PM_FW_DL_ERR)))
|
|
#define MHI_PM_IN_ERROR_STATE(pm_state) (pm_state >= MHI_PM_FW_DL_ERR)
|
|
#define MHI_PM_IN_FATAL_STATE(pm_state) (pm_state == MHI_PM_LD_ERR_FATAL_DETECT)
|
|
#define MHI_DB_ACCESS_VALID(mhi_cntrl) (mhi_cntrl->pm_state & mhi_cntrl->db_access)
|
|
#define MHI_WAKE_DB_CLEAR_VALID(pm_state) (pm_state & (MHI_PM_M0 | \
|
|
MHI_PM_M2 | MHI_PM_M3_EXIT))
|
|
#define MHI_WAKE_DB_SET_VALID(pm_state) (pm_state & MHI_PM_M2)
|
|
#define MHI_WAKE_DB_FORCE_SET_VALID(pm_state) MHI_WAKE_DB_CLEAR_VALID(pm_state)
|
|
#define MHI_EVENT_ACCESS_INVALID(pm_state) (pm_state == MHI_PM_DISABLE || \
|
|
MHI_PM_IN_ERROR_STATE(pm_state))
|
|
#define MHI_PM_IN_SUSPEND_STATE(pm_state) (pm_state & \
|
|
(MHI_PM_M3_ENTER | MHI_PM_M3))
|
|
|
|
#define NR_OF_CMD_RINGS 1
|
|
#define CMD_EL_PER_RING 128
|
|
#define PRIMARY_CMD_RING 0
|
|
#define MHI_DEV_WAKE_DB 127
|
|
#define MHI_MAX_MTU 0xffff
|
|
#define MHI_RANDOM_U32_NONZERO(bmsk) (prandom_u32_max(bmsk) + 1)
|
|
|
|
enum mhi_er_type {
|
|
MHI_ER_TYPE_INVALID = 0x0,
|
|
MHI_ER_TYPE_VALID = 0x1,
|
|
};
|
|
|
|
struct db_cfg {
|
|
bool reset_req;
|
|
bool db_mode;
|
|
u32 pollcfg;
|
|
enum mhi_db_brst_mode brstmode;
|
|
dma_addr_t db_val;
|
|
void (*process_db)(struct mhi_controller *mhi_cntrl,
|
|
struct db_cfg *db_cfg, void __iomem *io_addr,
|
|
dma_addr_t db_val);
|
|
};
|
|
|
|
struct mhi_pm_transitions {
|
|
enum mhi_pm_state from_state;
|
|
u32 to_states;
|
|
};
|
|
|
|
struct state_transition {
|
|
struct list_head node;
|
|
enum dev_st_transition state;
|
|
};
|
|
|
|
struct mhi_ring {
|
|
dma_addr_t dma_handle;
|
|
dma_addr_t iommu_base;
|
|
__le64 *ctxt_wp; /* point to ctxt wp */
|
|
void *pre_aligned;
|
|
void *base;
|
|
void *rp;
|
|
void *wp;
|
|
size_t el_size;
|
|
size_t len;
|
|
size_t elements;
|
|
size_t alloc_size;
|
|
void __iomem *db_addr;
|
|
};
|
|
|
|
struct mhi_cmd {
|
|
struct mhi_ring ring;
|
|
spinlock_t lock;
|
|
};
|
|
|
|
struct mhi_buf_info {
|
|
void *v_addr;
|
|
void *bb_addr;
|
|
void *wp;
|
|
void *cb_buf;
|
|
dma_addr_t p_addr;
|
|
size_t len;
|
|
enum dma_data_direction dir;
|
|
bool used; /* Indicates whether the buffer is used or not */
|
|
bool pre_mapped; /* Already pre-mapped by client */
|
|
};
|
|
|
|
struct mhi_event {
|
|
struct mhi_controller *mhi_cntrl;
|
|
struct mhi_chan *mhi_chan; /* dedicated to channel */
|
|
u32 er_index;
|
|
u32 intmod;
|
|
u32 irq;
|
|
int chan; /* this event ring is dedicated to a channel (optional) */
|
|
u32 priority;
|
|
enum mhi_er_data_type data_type;
|
|
struct mhi_ring ring;
|
|
struct db_cfg db_cfg;
|
|
struct tasklet_struct task;
|
|
spinlock_t lock;
|
|
int (*process_event)(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_event *mhi_event,
|
|
u32 event_quota);
|
|
bool hw_ring;
|
|
bool cl_manage;
|
|
bool offload_ev; /* managed by a device driver */
|
|
};
|
|
|
|
struct mhi_chan {
|
|
const char *name;
|
|
/*
|
|
* Important: When consuming, increment tre_ring first and when
|
|
* releasing, decrement buf_ring first. If tre_ring has space, buf_ring
|
|
* is guranteed to have space so we do not need to check both rings.
|
|
*/
|
|
struct mhi_ring buf_ring;
|
|
struct mhi_ring tre_ring;
|
|
u32 chan;
|
|
u32 er_index;
|
|
u32 intmod;
|
|
enum mhi_ch_type type;
|
|
enum dma_data_direction dir;
|
|
struct db_cfg db_cfg;
|
|
enum mhi_ch_ee_mask ee_mask;
|
|
enum mhi_ch_state ch_state;
|
|
enum mhi_ev_ccs ccs;
|
|
struct mhi_device *mhi_dev;
|
|
void (*xfer_cb)(struct mhi_device *mhi_dev, struct mhi_result *result);
|
|
struct mutex mutex;
|
|
struct completion completion;
|
|
rwlock_t lock;
|
|
struct list_head node;
|
|
bool lpm_notify;
|
|
bool configured;
|
|
bool offload_ch;
|
|
bool pre_alloc;
|
|
bool wake_capable;
|
|
};
|
|
|
|
/* Default MHI timeout */
|
|
#define MHI_TIMEOUT_MS (1000)
|
|
|
|
/* debugfs related functions */
|
|
#ifdef CONFIG_MHI_BUS_DEBUG
|
|
void mhi_create_debugfs(struct mhi_controller *mhi_cntrl);
|
|
void mhi_destroy_debugfs(struct mhi_controller *mhi_cntrl);
|
|
void mhi_debugfs_init(void);
|
|
void mhi_debugfs_exit(void);
|
|
#else
|
|
static inline void mhi_create_debugfs(struct mhi_controller *mhi_cntrl)
|
|
{
|
|
}
|
|
|
|
static inline void mhi_destroy_debugfs(struct mhi_controller *mhi_cntrl)
|
|
{
|
|
}
|
|
|
|
static inline void mhi_debugfs_init(void)
|
|
{
|
|
}
|
|
|
|
static inline void mhi_debugfs_exit(void)
|
|
{
|
|
}
|
|
#endif
|
|
|
|
struct mhi_device *mhi_alloc_device(struct mhi_controller *mhi_cntrl);
|
|
|
|
int mhi_destroy_device(struct device *dev, void *data);
|
|
void mhi_create_devices(struct mhi_controller *mhi_cntrl);
|
|
|
|
int mhi_alloc_bhie_table(struct mhi_controller *mhi_cntrl,
|
|
struct image_info **image_info, size_t alloc_size);
|
|
void mhi_free_bhie_table(struct mhi_controller *mhi_cntrl,
|
|
struct image_info *image_info);
|
|
|
|
/* Power management APIs */
|
|
enum mhi_pm_state __must_check mhi_tryset_pm_state(
|
|
struct mhi_controller *mhi_cntrl,
|
|
enum mhi_pm_state state);
|
|
const char *to_mhi_pm_state_str(u32 state);
|
|
int mhi_queue_state_transition(struct mhi_controller *mhi_cntrl,
|
|
enum dev_st_transition state);
|
|
void mhi_pm_st_worker(struct work_struct *work);
|
|
void mhi_pm_sys_err_handler(struct mhi_controller *mhi_cntrl);
|
|
int mhi_ready_state_transition(struct mhi_controller *mhi_cntrl);
|
|
int mhi_pm_m0_transition(struct mhi_controller *mhi_cntrl);
|
|
void mhi_pm_m1_transition(struct mhi_controller *mhi_cntrl);
|
|
int mhi_pm_m3_transition(struct mhi_controller *mhi_cntrl);
|
|
int __mhi_device_get_sync(struct mhi_controller *mhi_cntrl);
|
|
int mhi_send_cmd(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan,
|
|
enum mhi_cmd_type cmd);
|
|
int mhi_download_amss_image(struct mhi_controller *mhi_cntrl);
|
|
static inline bool mhi_is_active(struct mhi_controller *mhi_cntrl)
|
|
{
|
|
return (mhi_cntrl->dev_state >= MHI_STATE_M0 &&
|
|
mhi_cntrl->dev_state <= MHI_STATE_M3_FAST);
|
|
}
|
|
|
|
static inline void mhi_trigger_resume(struct mhi_controller *mhi_cntrl)
|
|
{
|
|
pm_wakeup_event(&mhi_cntrl->mhi_dev->dev, 0);
|
|
mhi_cntrl->runtime_get(mhi_cntrl);
|
|
mhi_cntrl->runtime_put(mhi_cntrl);
|
|
}
|
|
|
|
/* Register access methods */
|
|
void mhi_db_brstmode(struct mhi_controller *mhi_cntrl, struct db_cfg *db_cfg,
|
|
void __iomem *db_addr, dma_addr_t db_val);
|
|
void mhi_db_brstmode_disable(struct mhi_controller *mhi_cntrl,
|
|
struct db_cfg *db_mode, void __iomem *db_addr,
|
|
dma_addr_t db_val);
|
|
int __must_check mhi_read_reg(struct mhi_controller *mhi_cntrl,
|
|
void __iomem *base, u32 offset, u32 *out);
|
|
int __must_check mhi_read_reg_field(struct mhi_controller *mhi_cntrl,
|
|
void __iomem *base, u32 offset, u32 mask,
|
|
u32 *out);
|
|
int __must_check mhi_poll_reg_field(struct mhi_controller *mhi_cntrl,
|
|
void __iomem *base, u32 offset, u32 mask,
|
|
u32 val, u32 delayus);
|
|
void mhi_write_reg(struct mhi_controller *mhi_cntrl, void __iomem *base,
|
|
u32 offset, u32 val);
|
|
int __must_check mhi_write_reg_field(struct mhi_controller *mhi_cntrl,
|
|
void __iomem *base, u32 offset, u32 mask,
|
|
u32 val);
|
|
void mhi_ring_er_db(struct mhi_event *mhi_event);
|
|
void mhi_write_db(struct mhi_controller *mhi_cntrl, void __iomem *db_addr,
|
|
dma_addr_t db_val);
|
|
void mhi_ring_cmd_db(struct mhi_controller *mhi_cntrl, struct mhi_cmd *mhi_cmd);
|
|
void mhi_ring_chan_db(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_chan *mhi_chan);
|
|
|
|
/* Initialization methods */
|
|
int mhi_init_mmio(struct mhi_controller *mhi_cntrl);
|
|
int mhi_init_dev_ctxt(struct mhi_controller *mhi_cntrl);
|
|
void mhi_deinit_dev_ctxt(struct mhi_controller *mhi_cntrl);
|
|
int mhi_init_irq_setup(struct mhi_controller *mhi_cntrl);
|
|
void mhi_deinit_free_irq(struct mhi_controller *mhi_cntrl);
|
|
int mhi_rddm_prepare(struct mhi_controller *mhi_cntrl,
|
|
struct image_info *img_info);
|
|
void mhi_fw_load_handler(struct mhi_controller *mhi_cntrl);
|
|
|
|
/* Automatically allocate and queue inbound buffers */
|
|
#define MHI_CH_INBOUND_ALLOC_BUFS BIT(0)
|
|
int mhi_prepare_channel(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_chan *mhi_chan, unsigned int flags);
|
|
|
|
int mhi_init_chan_ctxt(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_chan *mhi_chan);
|
|
void mhi_deinit_chan_ctxt(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_chan *mhi_chan);
|
|
void mhi_reset_chan(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_chan *mhi_chan);
|
|
|
|
/* Event processing methods */
|
|
void mhi_ctrl_ev_task(unsigned long data);
|
|
void mhi_ev_task(unsigned long data);
|
|
int mhi_process_data_event_ring(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_event *mhi_event, u32 event_quota);
|
|
int mhi_process_ctrl_ev_ring(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_event *mhi_event, u32 event_quota);
|
|
|
|
/* ISR handlers */
|
|
irqreturn_t mhi_irq_handler(int irq_number, void *dev);
|
|
irqreturn_t mhi_intvec_threaded_handler(int irq_number, void *dev);
|
|
irqreturn_t mhi_intvec_handler(int irq_number, void *dev);
|
|
|
|
int mhi_gen_tre(struct mhi_controller *mhi_cntrl, struct mhi_chan *mhi_chan,
|
|
struct mhi_buf_info *info, enum mhi_flags flags);
|
|
int mhi_map_single_no_bb(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_buf_info *buf_info);
|
|
int mhi_map_single_use_bb(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_buf_info *buf_info);
|
|
void mhi_unmap_single_no_bb(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_buf_info *buf_info);
|
|
void mhi_unmap_single_use_bb(struct mhi_controller *mhi_cntrl,
|
|
struct mhi_buf_info *buf_info);
|
|
|
|
#endif /* _MHI_INT_H */
|