Path: blob/master/drivers/crypto/cavium/cpt/cptvf_main.c
51980 views
// SPDX-License-Identifier: GPL-2.0-only1/*2* Copyright (C) 2016 Cavium, Inc.3*/45#include <linux/interrupt.h>6#include <linux/module.h>78#include "cptvf.h"910#define DRV_NAME "thunder-cptvf"11#define DRV_VERSION "1.0"1213struct cptvf_wqe {14struct tasklet_struct twork;15void *cptvf;16u32 qno;17};1819struct cptvf_wqe_info {20struct cptvf_wqe vq_wqe[CPT_NUM_QS_PER_VF];21};2223static void vq_work_handler(unsigned long data)24{25struct cptvf_wqe_info *cwqe_info = (struct cptvf_wqe_info *)data;26struct cptvf_wqe *cwqe = &cwqe_info->vq_wqe[0];2728vq_post_process(cwqe->cptvf, cwqe->qno);29}3031static int init_worker_threads(struct cpt_vf *cptvf)32{33struct pci_dev *pdev = cptvf->pdev;34struct cptvf_wqe_info *cwqe_info;35int i;3637cwqe_info = kzalloc(sizeof(*cwqe_info), GFP_KERNEL);38if (!cwqe_info)39return -ENOMEM;4041if (cptvf->nr_queues) {42dev_info(&pdev->dev, "Creating VQ worker threads (%d)\n",43cptvf->nr_queues);44}4546for (i = 0; i < cptvf->nr_queues; i++) {47tasklet_init(&cwqe_info->vq_wqe[i].twork, vq_work_handler,48(u64)cwqe_info);49cwqe_info->vq_wqe[i].qno = i;50cwqe_info->vq_wqe[i].cptvf = cptvf;51}5253cptvf->wqe_info = cwqe_info;5455return 0;56}5758static void cleanup_worker_threads(struct cpt_vf *cptvf)59{60struct cptvf_wqe_info *cwqe_info;61struct pci_dev *pdev = cptvf->pdev;62int i;6364cwqe_info = (struct cptvf_wqe_info *)cptvf->wqe_info;65if (!cwqe_info)66return;6768if (cptvf->nr_queues) {69dev_info(&pdev->dev, "Cleaning VQ worker threads (%u)\n",70cptvf->nr_queues);71}7273for (i = 0; i < cptvf->nr_queues; i++)74tasklet_kill(&cwqe_info->vq_wqe[i].twork);7576kfree_sensitive(cwqe_info);77cptvf->wqe_info = NULL;78}7980static void free_pending_queues(struct pending_qinfo *pqinfo)81{82int i;83struct pending_queue *queue;8485for_each_pending_queue(pqinfo, queue, i) {86if (!queue->head)87continue;8889/* free single queue */90kfree_sensitive((queue->head));9192queue->front = 0;93queue->rear = 0;9495return;96}9798pqinfo->qlen = 0;99pqinfo->nr_queues = 0;100}101102static int alloc_pending_queues(struct pending_qinfo *pqinfo, u32 qlen,103u32 nr_queues)104{105u32 i;106int ret;107struct pending_queue *queue = NULL;108109pqinfo->nr_queues = nr_queues;110pqinfo->qlen = qlen;111112for_each_pending_queue(pqinfo, queue, i) {113queue->head = kcalloc(qlen, sizeof(*queue->head), GFP_KERNEL);114if (!queue->head) {115ret = -ENOMEM;116goto pending_qfail;117}118119queue->front = 0;120queue->rear = 0;121atomic64_set((&queue->pending_count), (0));122123/* init queue spin lock */124spin_lock_init(&queue->lock);125}126127return 0;128129pending_qfail:130free_pending_queues(pqinfo);131132return ret;133}134135static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues)136{137struct pci_dev *pdev = cptvf->pdev;138int ret;139140if (!nr_queues)141return 0;142143ret = alloc_pending_queues(&cptvf->pqinfo, qlen, nr_queues);144if (ret) {145dev_err(&pdev->dev, "failed to setup pending queues (%u)\n",146nr_queues);147return ret;148}149150return 0;151}152153static void cleanup_pending_queues(struct cpt_vf *cptvf)154{155struct pci_dev *pdev = cptvf->pdev;156157if (!cptvf->nr_queues)158return;159160dev_info(&pdev->dev, "Cleaning VQ pending queue (%u)\n",161cptvf->nr_queues);162free_pending_queues(&cptvf->pqinfo);163}164165static void free_command_queues(struct cpt_vf *cptvf,166struct command_qinfo *cqinfo)167{168int i;169struct command_queue *queue = NULL;170struct command_chunk *chunk = NULL;171struct pci_dev *pdev = cptvf->pdev;172struct hlist_node *node;173174/* clean up for each queue */175for (i = 0; i < cptvf->nr_queues; i++) {176queue = &cqinfo->queue[i];177if (hlist_empty(&cqinfo->queue[i].chead))178continue;179180hlist_for_each_entry_safe(chunk, node, &cqinfo->queue[i].chead,181nextchunk) {182dma_free_coherent(&pdev->dev,183chunk->size + CPT_NEXT_CHUNK_PTR_SIZE,184chunk->head,185chunk->dma_addr);186chunk->head = NULL;187chunk->dma_addr = 0;188hlist_del(&chunk->nextchunk);189kfree_sensitive(chunk);190}191192queue->nchunks = 0;193queue->idx = 0;194}195196/* common cleanup */197cqinfo->cmd_size = 0;198}199200static int alloc_command_queues(struct cpt_vf *cptvf,201struct command_qinfo *cqinfo, size_t cmd_size,202u32 qlen)203{204int i;205size_t q_size;206struct command_queue *queue = NULL;207struct pci_dev *pdev = cptvf->pdev;208209/* common init */210cqinfo->cmd_size = cmd_size;211/* Qsize in dwords, needed for SADDR config, 1-next chunk pointer */212cptvf->qsize = min(qlen, cqinfo->qchunksize) *213CPT_NEXT_CHUNK_PTR_SIZE + 1;214/* Qsize in bytes to create space for alignment */215q_size = qlen * cqinfo->cmd_size;216217/* per queue initialization */218for (i = 0; i < cptvf->nr_queues; i++) {219size_t c_size = 0;220size_t rem_q_size = q_size;221struct command_chunk *curr = NULL, *first = NULL, *last = NULL;222u32 qcsize_bytes = cqinfo->qchunksize * cqinfo->cmd_size;223224queue = &cqinfo->queue[i];225INIT_HLIST_HEAD(&cqinfo->queue[i].chead);226do {227curr = kzalloc(sizeof(*curr), GFP_KERNEL);228if (!curr)229goto cmd_qfail;230231c_size = (rem_q_size > qcsize_bytes) ? qcsize_bytes :232rem_q_size;233curr->head = dma_alloc_coherent(&pdev->dev,234c_size + CPT_NEXT_CHUNK_PTR_SIZE,235&curr->dma_addr,236GFP_KERNEL);237if (!curr->head) {238dev_err(&pdev->dev, "Command Q (%d) chunk (%d) allocation failed\n",239i, queue->nchunks);240kfree(curr);241goto cmd_qfail;242}243244curr->size = c_size;245if (queue->nchunks == 0) {246hlist_add_head(&curr->nextchunk,247&cqinfo->queue[i].chead);248first = curr;249} else {250hlist_add_behind(&curr->nextchunk,251&last->nextchunk);252}253254queue->nchunks++;255rem_q_size -= c_size;256if (last)257*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;258259last = curr;260} while (rem_q_size);261262/* Make the queue circular */263/* Tie back last chunk entry to head */264curr = first;265*((u64 *)(&last->head[last->size])) = (u64)curr->dma_addr;266queue->qhead = curr;267spin_lock_init(&queue->lock);268}269return 0;270271cmd_qfail:272free_command_queues(cptvf, cqinfo);273return -ENOMEM;274}275276static int init_command_queues(struct cpt_vf *cptvf, u32 qlen)277{278struct pci_dev *pdev = cptvf->pdev;279int ret;280281/* setup AE command queues */282ret = alloc_command_queues(cptvf, &cptvf->cqinfo, CPT_INST_SIZE,283qlen);284if (ret) {285dev_err(&pdev->dev, "failed to allocate AE command queues (%u)\n",286cptvf->nr_queues);287return ret;288}289290return ret;291}292293static void cleanup_command_queues(struct cpt_vf *cptvf)294{295struct pci_dev *pdev = cptvf->pdev;296297if (!cptvf->nr_queues)298return;299300dev_info(&pdev->dev, "Cleaning VQ command queue (%u)\n",301cptvf->nr_queues);302free_command_queues(cptvf, &cptvf->cqinfo);303}304305static void cptvf_sw_cleanup(struct cpt_vf *cptvf)306{307cleanup_worker_threads(cptvf);308cleanup_pending_queues(cptvf);309cleanup_command_queues(cptvf);310}311312static int cptvf_sw_init(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues)313{314struct pci_dev *pdev = cptvf->pdev;315int ret = 0;316u32 max_dev_queues = 0;317318max_dev_queues = CPT_NUM_QS_PER_VF;319/* possible cpus */320nr_queues = min_t(u32, nr_queues, max_dev_queues);321cptvf->nr_queues = nr_queues;322323ret = init_command_queues(cptvf, qlen);324if (ret) {325dev_err(&pdev->dev, "Failed to setup command queues (%u)\n",326nr_queues);327return ret;328}329330ret = init_pending_queues(cptvf, qlen, nr_queues);331if (ret) {332dev_err(&pdev->dev, "Failed to setup pending queues (%u)\n",333nr_queues);334goto setup_pqfail;335}336337/* Create worker threads for BH processing */338ret = init_worker_threads(cptvf);339if (ret) {340dev_err(&pdev->dev, "Failed to setup worker threads\n");341goto init_work_fail;342}343344return 0;345346init_work_fail:347cleanup_worker_threads(cptvf);348cleanup_pending_queues(cptvf);349350setup_pqfail:351cleanup_command_queues(cptvf);352353return ret;354}355356static void cptvf_free_irq_affinity(struct cpt_vf *cptvf, int vec)357{358irq_set_affinity_hint(pci_irq_vector(cptvf->pdev, vec), NULL);359free_cpumask_var(cptvf->affinity_mask[vec]);360}361362static void cptvf_write_vq_ctl(struct cpt_vf *cptvf, bool val)363{364union cptx_vqx_ctl vqx_ctl;365366vqx_ctl.u = cpt_read_csr64(cptvf->reg_base, CPTX_VQX_CTL(0, 0));367vqx_ctl.s.ena = val;368cpt_write_csr64(cptvf->reg_base, CPTX_VQX_CTL(0, 0), vqx_ctl.u);369}370371void cptvf_write_vq_doorbell(struct cpt_vf *cptvf, u32 val)372{373union cptx_vqx_doorbell vqx_dbell;374375vqx_dbell.u = cpt_read_csr64(cptvf->reg_base,376CPTX_VQX_DOORBELL(0, 0));377vqx_dbell.s.dbell_cnt = val * 8; /* Num of Instructions * 8 words */378cpt_write_csr64(cptvf->reg_base, CPTX_VQX_DOORBELL(0, 0),379vqx_dbell.u);380}381382static void cptvf_write_vq_inprog(struct cpt_vf *cptvf, u8 val)383{384union cptx_vqx_inprog vqx_inprg;385386vqx_inprg.u = cpt_read_csr64(cptvf->reg_base, CPTX_VQX_INPROG(0, 0));387vqx_inprg.s.inflight = val;388cpt_write_csr64(cptvf->reg_base, CPTX_VQX_INPROG(0, 0), vqx_inprg.u);389}390391static void cptvf_write_vq_done_numwait(struct cpt_vf *cptvf, u32 val)392{393union cptx_vqx_done_wait vqx_dwait;394395vqx_dwait.u = cpt_read_csr64(cptvf->reg_base,396CPTX_VQX_DONE_WAIT(0, 0));397vqx_dwait.s.num_wait = val;398cpt_write_csr64(cptvf->reg_base, CPTX_VQX_DONE_WAIT(0, 0),399vqx_dwait.u);400}401402static void cptvf_write_vq_done_timewait(struct cpt_vf *cptvf, u16 time)403{404union cptx_vqx_done_wait vqx_dwait;405406vqx_dwait.u = cpt_read_csr64(cptvf->reg_base,407CPTX_VQX_DONE_WAIT(0, 0));408vqx_dwait.s.time_wait = time;409cpt_write_csr64(cptvf->reg_base, CPTX_VQX_DONE_WAIT(0, 0),410vqx_dwait.u);411}412413static void cptvf_enable_swerr_interrupts(struct cpt_vf *cptvf)414{415union cptx_vqx_misc_ena_w1s vqx_misc_ena;416417vqx_misc_ena.u = cpt_read_csr64(cptvf->reg_base,418CPTX_VQX_MISC_ENA_W1S(0, 0));419/* Set mbox(0) interupts for the requested vf */420vqx_misc_ena.s.swerr = 1;421cpt_write_csr64(cptvf->reg_base, CPTX_VQX_MISC_ENA_W1S(0, 0),422vqx_misc_ena.u);423}424425static void cptvf_enable_mbox_interrupts(struct cpt_vf *cptvf)426{427union cptx_vqx_misc_ena_w1s vqx_misc_ena;428429vqx_misc_ena.u = cpt_read_csr64(cptvf->reg_base,430CPTX_VQX_MISC_ENA_W1S(0, 0));431/* Set mbox(0) interupts for the requested vf */432vqx_misc_ena.s.mbox = 1;433cpt_write_csr64(cptvf->reg_base, CPTX_VQX_MISC_ENA_W1S(0, 0),434vqx_misc_ena.u);435}436437static void cptvf_enable_done_interrupts(struct cpt_vf *cptvf)438{439union cptx_vqx_done_ena_w1s vqx_done_ena;440441vqx_done_ena.u = cpt_read_csr64(cptvf->reg_base,442CPTX_VQX_DONE_ENA_W1S(0, 0));443/* Set DONE interrupt for the requested vf */444vqx_done_ena.s.done = 1;445cpt_write_csr64(cptvf->reg_base, CPTX_VQX_DONE_ENA_W1S(0, 0),446vqx_done_ena.u);447}448449static void cptvf_clear_dovf_intr(struct cpt_vf *cptvf)450{451union cptx_vqx_misc_int vqx_misc_int;452453vqx_misc_int.u = cpt_read_csr64(cptvf->reg_base,454CPTX_VQX_MISC_INT(0, 0));455/* W1C for the VF */456vqx_misc_int.s.dovf = 1;457cpt_write_csr64(cptvf->reg_base, CPTX_VQX_MISC_INT(0, 0),458vqx_misc_int.u);459}460461static void cptvf_clear_irde_intr(struct cpt_vf *cptvf)462{463union cptx_vqx_misc_int vqx_misc_int;464465vqx_misc_int.u = cpt_read_csr64(cptvf->reg_base,466CPTX_VQX_MISC_INT(0, 0));467/* W1C for the VF */468vqx_misc_int.s.irde = 1;469cpt_write_csr64(cptvf->reg_base, CPTX_VQX_MISC_INT(0, 0),470vqx_misc_int.u);471}472473static void cptvf_clear_nwrp_intr(struct cpt_vf *cptvf)474{475union cptx_vqx_misc_int vqx_misc_int;476477vqx_misc_int.u = cpt_read_csr64(cptvf->reg_base,478CPTX_VQX_MISC_INT(0, 0));479/* W1C for the VF */480vqx_misc_int.s.nwrp = 1;481cpt_write_csr64(cptvf->reg_base,482CPTX_VQX_MISC_INT(0, 0), vqx_misc_int.u);483}484485static void cptvf_clear_mbox_intr(struct cpt_vf *cptvf)486{487union cptx_vqx_misc_int vqx_misc_int;488489vqx_misc_int.u = cpt_read_csr64(cptvf->reg_base,490CPTX_VQX_MISC_INT(0, 0));491/* W1C for the VF */492vqx_misc_int.s.mbox = 1;493cpt_write_csr64(cptvf->reg_base, CPTX_VQX_MISC_INT(0, 0),494vqx_misc_int.u);495}496497static void cptvf_clear_swerr_intr(struct cpt_vf *cptvf)498{499union cptx_vqx_misc_int vqx_misc_int;500501vqx_misc_int.u = cpt_read_csr64(cptvf->reg_base,502CPTX_VQX_MISC_INT(0, 0));503/* W1C for the VF */504vqx_misc_int.s.swerr = 1;505cpt_write_csr64(cptvf->reg_base, CPTX_VQX_MISC_INT(0, 0),506vqx_misc_int.u);507}508509static u64 cptvf_read_vf_misc_intr_status(struct cpt_vf *cptvf)510{511return cpt_read_csr64(cptvf->reg_base, CPTX_VQX_MISC_INT(0, 0));512}513514static irqreturn_t cptvf_misc_intr_handler(int irq, void *cptvf_irq)515{516struct cpt_vf *cptvf = (struct cpt_vf *)cptvf_irq;517struct pci_dev *pdev = cptvf->pdev;518u64 intr;519520intr = cptvf_read_vf_misc_intr_status(cptvf);521/*Check for MISC interrupt types*/522if (likely(intr & CPT_VF_INTR_MBOX_MASK)) {523dev_dbg(&pdev->dev, "Mailbox interrupt 0x%llx on CPT VF %d\n",524intr, cptvf->vfid);525cptvf_handle_mbox_intr(cptvf);526cptvf_clear_mbox_intr(cptvf);527} else if (unlikely(intr & CPT_VF_INTR_DOVF_MASK)) {528cptvf_clear_dovf_intr(cptvf);529/*Clear doorbell count*/530cptvf_write_vq_doorbell(cptvf, 0);531dev_err(&pdev->dev, "Doorbell overflow error interrupt 0x%llx on CPT VF %d\n",532intr, cptvf->vfid);533} else if (unlikely(intr & CPT_VF_INTR_IRDE_MASK)) {534cptvf_clear_irde_intr(cptvf);535dev_err(&pdev->dev, "Instruction NCB read error interrupt 0x%llx on CPT VF %d\n",536intr, cptvf->vfid);537} else if (unlikely(intr & CPT_VF_INTR_NWRP_MASK)) {538cptvf_clear_nwrp_intr(cptvf);539dev_err(&pdev->dev, "NCB response write error interrupt 0x%llx on CPT VF %d\n",540intr, cptvf->vfid);541} else if (unlikely(intr & CPT_VF_INTR_SERR_MASK)) {542cptvf_clear_swerr_intr(cptvf);543dev_err(&pdev->dev, "Software error interrupt 0x%llx on CPT VF %d\n",544intr, cptvf->vfid);545} else {546dev_err(&pdev->dev, "Unhandled interrupt in CPT VF %d\n",547cptvf->vfid);548}549550return IRQ_HANDLED;551}552553static inline struct cptvf_wqe *get_cptvf_vq_wqe(struct cpt_vf *cptvf,554int qno)555{556struct cptvf_wqe_info *nwqe_info;557558if (unlikely(qno >= cptvf->nr_queues))559return NULL;560nwqe_info = (struct cptvf_wqe_info *)cptvf->wqe_info;561562return &nwqe_info->vq_wqe[qno];563}564565static inline u32 cptvf_read_vq_done_count(struct cpt_vf *cptvf)566{567union cptx_vqx_done vqx_done;568569vqx_done.u = cpt_read_csr64(cptvf->reg_base, CPTX_VQX_DONE(0, 0));570return vqx_done.s.done;571}572573static inline void cptvf_write_vq_done_ack(struct cpt_vf *cptvf,574u32 ackcnt)575{576union cptx_vqx_done_ack vqx_dack_cnt;577578vqx_dack_cnt.u = cpt_read_csr64(cptvf->reg_base,579CPTX_VQX_DONE_ACK(0, 0));580vqx_dack_cnt.s.done_ack = ackcnt;581cpt_write_csr64(cptvf->reg_base, CPTX_VQX_DONE_ACK(0, 0),582vqx_dack_cnt.u);583}584585static irqreturn_t cptvf_done_intr_handler(int irq, void *cptvf_irq)586{587struct cpt_vf *cptvf = (struct cpt_vf *)cptvf_irq;588struct pci_dev *pdev = cptvf->pdev;589/* Read the number of completions */590u32 intr = cptvf_read_vq_done_count(cptvf);591592if (intr) {593struct cptvf_wqe *wqe;594595/* Acknowledge the number of596* scheduled completions for processing597*/598cptvf_write_vq_done_ack(cptvf, intr);599wqe = get_cptvf_vq_wqe(cptvf, 0);600if (unlikely(!wqe)) {601dev_err(&pdev->dev, "No work to schedule for VF (%d)",602cptvf->vfid);603return IRQ_NONE;604}605tasklet_hi_schedule(&wqe->twork);606}607608return IRQ_HANDLED;609}610611static void cptvf_set_irq_affinity(struct cpt_vf *cptvf, int vec)612{613struct pci_dev *pdev = cptvf->pdev;614int cpu;615616if (!zalloc_cpumask_var(&cptvf->affinity_mask[vec],617GFP_KERNEL)) {618dev_err(&pdev->dev, "Allocation failed for affinity_mask for VF %d",619cptvf->vfid);620return;621}622623cpu = cptvf->vfid % num_online_cpus();624cpumask_set_cpu(cpumask_local_spread(cpu, cptvf->node),625cptvf->affinity_mask[vec]);626irq_set_affinity_hint(pci_irq_vector(pdev, vec),627cptvf->affinity_mask[vec]);628}629630static void cptvf_write_vq_saddr(struct cpt_vf *cptvf, u64 val)631{632union cptx_vqx_saddr vqx_saddr;633634vqx_saddr.u = val;635cpt_write_csr64(cptvf->reg_base, CPTX_VQX_SADDR(0, 0), vqx_saddr.u);636}637638static void cptvf_device_init(struct cpt_vf *cptvf)639{640u64 base_addr = 0;641642/* Disable the VQ */643cptvf_write_vq_ctl(cptvf, 0);644/* Reset the doorbell */645cptvf_write_vq_doorbell(cptvf, 0);646/* Clear inflight */647cptvf_write_vq_inprog(cptvf, 0);648/* Write VQ SADDR */649/* TODO: for now only one queue, so hard coded */650base_addr = (u64)(cptvf->cqinfo.queue[0].qhead->dma_addr);651cptvf_write_vq_saddr(cptvf, base_addr);652/* Configure timerhold / coalescence */653cptvf_write_vq_done_timewait(cptvf, CPT_TIMER_THOLD);654cptvf_write_vq_done_numwait(cptvf, 1);655/* Enable the VQ */656cptvf_write_vq_ctl(cptvf, 1);657/* Flag the VF ready */658cptvf->flags |= CPT_FLAG_DEVICE_READY;659}660661static int cptvf_probe(struct pci_dev *pdev, const struct pci_device_id *ent)662{663struct device *dev = &pdev->dev;664struct cpt_vf *cptvf;665int err;666667cptvf = devm_kzalloc(dev, sizeof(*cptvf), GFP_KERNEL);668if (!cptvf)669return -ENOMEM;670671pci_set_drvdata(pdev, cptvf);672cptvf->pdev = pdev;673err = pci_enable_device(pdev);674if (err) {675dev_err(dev, "Failed to enable PCI device\n");676pci_set_drvdata(pdev, NULL);677return err;678}679680err = pci_request_regions(pdev, DRV_NAME);681if (err) {682dev_err(dev, "PCI request regions failed 0x%x\n", err);683goto cptvf_err_disable_device;684}685/* Mark as VF driver */686cptvf->flags |= CPT_FLAG_VF_DRIVER;687err = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(48));688if (err) {689dev_err(dev, "Unable to get usable 48-bit DMA configuration\n");690goto cptvf_err_release_regions;691}692693/* MAP PF's configuration registers */694cptvf->reg_base = pcim_iomap(pdev, 0, 0);695if (!cptvf->reg_base) {696dev_err(dev, "Cannot map config register space, aborting\n");697err = -ENOMEM;698goto cptvf_err_release_regions;699}700701cptvf->node = dev_to_node(&pdev->dev);702err = pci_alloc_irq_vectors(pdev, CPT_VF_MSIX_VECTORS,703CPT_VF_MSIX_VECTORS, PCI_IRQ_MSIX);704if (err < 0) {705dev_err(dev, "Request for #%d msix vectors failed\n",706CPT_VF_MSIX_VECTORS);707goto cptvf_err_release_regions;708}709710err = request_irq(pci_irq_vector(pdev, CPT_VF_INT_VEC_E_MISC),711cptvf_misc_intr_handler, 0, "CPT VF misc intr",712cptvf);713if (err) {714dev_err(dev, "Request misc irq failed");715goto cptvf_free_vectors;716}717718/* Enable mailbox interrupt */719cptvf_enable_mbox_interrupts(cptvf);720cptvf_enable_swerr_interrupts(cptvf);721722/* Check ready with PF */723/* Gets chip ID / device Id from PF if ready */724err = cptvf_check_pf_ready(cptvf);725if (err) {726dev_err(dev, "PF not responding to READY msg");727goto cptvf_free_misc_irq;728}729730/* CPT VF software resources initialization */731cptvf->cqinfo.qchunksize = CPT_CMD_QCHUNK_SIZE;732err = cptvf_sw_init(cptvf, CPT_CMD_QLEN, CPT_NUM_QS_PER_VF);733if (err) {734dev_err(dev, "cptvf_sw_init() failed");735goto cptvf_free_misc_irq;736}737/* Convey VQ LEN to PF */738err = cptvf_send_vq_size_msg(cptvf);739if (err) {740dev_err(dev, "PF not responding to QLEN msg");741goto cptvf_free_misc_irq;742}743744/* CPT VF device initialization */745cptvf_device_init(cptvf);746/* Send msg to PF to assign currnet Q to required group */747cptvf->vfgrp = 1;748err = cptvf_send_vf_to_grp_msg(cptvf);749if (err) {750dev_err(dev, "PF not responding to VF_GRP msg");751goto cptvf_free_misc_irq;752}753754cptvf->priority = 1;755err = cptvf_send_vf_priority_msg(cptvf);756if (err) {757dev_err(dev, "PF not responding to VF_PRIO msg");758goto cptvf_free_misc_irq;759}760761err = request_irq(pci_irq_vector(pdev, CPT_VF_INT_VEC_E_DONE),762cptvf_done_intr_handler, 0, "CPT VF done intr",763cptvf);764if (err) {765dev_err(dev, "Request done irq failed\n");766goto cptvf_free_misc_irq;767}768769/* Enable mailbox interrupt */770cptvf_enable_done_interrupts(cptvf);771772/* Set irq affinity masks */773cptvf_set_irq_affinity(cptvf, CPT_VF_INT_VEC_E_MISC);774cptvf_set_irq_affinity(cptvf, CPT_VF_INT_VEC_E_DONE);775776err = cptvf_send_vf_up(cptvf);777if (err) {778dev_err(dev, "PF not responding to UP msg");779goto cptvf_free_irq_affinity;780}781err = cvm_crypto_init(cptvf);782if (err) {783dev_err(dev, "Algorithm register failed\n");784goto cptvf_free_irq_affinity;785}786return 0;787788cptvf_free_irq_affinity:789cptvf_free_irq_affinity(cptvf, CPT_VF_INT_VEC_E_DONE);790cptvf_free_irq_affinity(cptvf, CPT_VF_INT_VEC_E_MISC);791cptvf_free_misc_irq:792free_irq(pci_irq_vector(pdev, CPT_VF_INT_VEC_E_MISC), cptvf);793cptvf_free_vectors:794pci_free_irq_vectors(cptvf->pdev);795cptvf_err_release_regions:796pci_release_regions(pdev);797cptvf_err_disable_device:798pci_disable_device(pdev);799pci_set_drvdata(pdev, NULL);800801return err;802}803804static void cptvf_remove(struct pci_dev *pdev)805{806struct cpt_vf *cptvf = pci_get_drvdata(pdev);807808if (!cptvf) {809dev_err(&pdev->dev, "Invalid CPT-VF device\n");810return;811}812813/* Convey DOWN to PF */814if (cptvf_send_vf_down(cptvf)) {815dev_err(&pdev->dev, "PF not responding to DOWN msg");816} else {817cptvf_free_irq_affinity(cptvf, CPT_VF_INT_VEC_E_DONE);818cptvf_free_irq_affinity(cptvf, CPT_VF_INT_VEC_E_MISC);819free_irq(pci_irq_vector(pdev, CPT_VF_INT_VEC_E_DONE), cptvf);820free_irq(pci_irq_vector(pdev, CPT_VF_INT_VEC_E_MISC), cptvf);821pci_free_irq_vectors(cptvf->pdev);822cptvf_sw_cleanup(cptvf);823pci_set_drvdata(pdev, NULL);824pci_release_regions(pdev);825pci_disable_device(pdev);826cvm_crypto_exit();827}828}829830static void cptvf_shutdown(struct pci_dev *pdev)831{832cptvf_remove(pdev);833}834835/* Supported devices */836static const struct pci_device_id cptvf_id_table[] = {837{PCI_VDEVICE(CAVIUM, CPT_81XX_PCI_VF_DEVICE_ID), 0},838{ 0, } /* end of table */839};840841static struct pci_driver cptvf_pci_driver = {842.name = DRV_NAME,843.id_table = cptvf_id_table,844.probe = cptvf_probe,845.remove = cptvf_remove,846.shutdown = cptvf_shutdown,847};848849module_pci_driver(cptvf_pci_driver);850851MODULE_AUTHOR("George Cherian <[email protected]>");852MODULE_DESCRIPTION("Cavium Thunder CPT Virtual Function Driver");853MODULE_LICENSE("GPL v2");854MODULE_VERSION(DRV_VERSION);855MODULE_DEVICE_TABLE(pci, cptvf_id_table);856857858