Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 494f401b authored by Suganath Prabu Subramani's avatar Suganath Prabu Subramani Committed by Martin K. Petersen
Browse files

scsi: mpt3sas: Fix sparse warnings



1) Used variable __le64/__le32 whichever required in building NVME
PRP, and passed to LE Controller.

2) Remove unused functions, And Declared functions as static which are
used only in mpt3sas_scsih.c.

Signed-off-by: default avatarChaitra P B <chaitra.basappa@broadcom.com>
Signed-off-by: default avatarSuganath Prabu S <suganath-prabu.subramani@broadcom.com>
Signed-off-by: default avatarMartin K. Petersen <martin.petersen@oracle.com>
parent cd5897ed
Loading
Loading
Loading
Loading
+11 −11
Original line number Diff line number Diff line
@@ -1437,8 +1437,8 @@ _base_build_nvme_prp(struct MPT3SAS_ADAPTER *ioc, u16 smid,
	size_t data_in_sz)
{
	int		prp_size = NVME_PRP_SIZE;
	u64		*prp_entry, *prp1_entry, *prp2_entry, *prp_entry_phys;
	u64		*prp_page, *prp_page_phys;
	__le64		*prp_entry, *prp1_entry, *prp2_entry, *prp_entry_phys;
	__le64		*prp_page, *prp_page_phys;
	u32		offset, entry_len;
	u32		page_mask_result, page_mask;
	dma_addr_t	paddr;
@@ -1455,17 +1455,17 @@ _base_build_nvme_prp(struct MPT3SAS_ADAPTER *ioc, u16 smid,
	 * PRP1 is located at a 24 byte offset from the start of the NVMe
	 * command.  Then set the current PRP entry pointer to PRP1.
	 */
	prp1_entry = (u64 *)(nvme_encap_request->NVMe_Command +
	prp1_entry = (__le64 *)(nvme_encap_request->NVMe_Command +
	    NVME_CMD_PRP1_OFFSET);
	prp2_entry = (u64 *)(nvme_encap_request->NVMe_Command +
	prp2_entry = (__le64 *)(nvme_encap_request->NVMe_Command +
	    NVME_CMD_PRP2_OFFSET);
	prp_entry = prp1_entry;
	/*
	 * For the PRP entries, use the specially allocated buffer of
	 * contiguous memory.
	 */
	prp_page = (u64 *)mpt3sas_base_get_pcie_sgl(ioc, smid);
	prp_page_phys = (u64 *)mpt3sas_base_get_pcie_sgl_dma(ioc, smid);
	prp_page = (__le64 *)mpt3sas_base_get_pcie_sgl(ioc, smid);
	prp_page_phys = (__le64 *)mpt3sas_base_get_pcie_sgl_dma(ioc, smid);

	/*
	 * Check if we are within 1 entry of a page boundary we don't
@@ -1475,8 +1475,8 @@ _base_build_nvme_prp(struct MPT3SAS_ADAPTER *ioc, u16 smid,
	page_mask_result = (uintptr_t)((u8 *)prp_page + prp_size) & page_mask;
	if (!page_mask_result) {
		/* Bump up to next page boundary. */
		prp_page = (u64 *)((u8 *)prp_page + prp_size);
		prp_page_phys = (u64 *)((u8 *)prp_page_phys + prp_size);
		prp_page = (__le64 *)((u8 *)prp_page + prp_size);
		prp_page_phys = (__le64 *)((u8 *)prp_page_phys + prp_size);
	}

	/*
@@ -1604,7 +1604,7 @@ _base_build_nvme_prp(struct MPT3SAS_ADAPTER *ioc, u16 smid,
 * Returns:		true: PRPs are built
 *			false: IEEE SGLs needs to be built
 */
void
static void
base_make_prp_nvme(struct MPT3SAS_ADAPTER *ioc,
		struct scsi_cmnd *scmd,
		Mpi25SCSIIORequest_t *mpi_request,
@@ -1612,7 +1612,7 @@ base_make_prp_nvme(struct MPT3SAS_ADAPTER *ioc,
{
	int sge_len, offset, num_prp_in_chain = 0;
	Mpi25IeeeSgeChain64_t *main_chain_element, *ptr_first_sgl;
	u64 *curr_buff;
	__le64 *curr_buff;
	dma_addr_t msg_phys;
	u64 sge_addr;
	u32 page_mask, page_mask_result;
@@ -1740,7 +1740,7 @@ base_is_prp_possible(struct MPT3SAS_ADAPTER *ioc,
	struct scatterlist *sg_scmd;
	bool build_prp = true;

	data_length = cpu_to_le32(scsi_bufflen(scmd));
	data_length = scsi_bufflen(scmd);
	sg_scmd = scsi_sglist(scmd);

	/* If Datalenth is <= 16K and number of SGE’s entries are <= 2
+5 −32
Original line number Diff line number Diff line
@@ -599,7 +599,7 @@ __mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc,
 *
 * This searches for pcie_device from target, then return pcie_device object.
 */
struct _pcie_device *
static struct _pcie_device *
mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc,
	struct MPT3SAS_TARGET *tgt_priv)
{
@@ -942,7 +942,7 @@ _scsih_sas_device_init_add(struct MPT3SAS_ADAPTER *ioc,
}


struct _pcie_device *
static struct _pcie_device *
__mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
{
	struct _pcie_device *pcie_device;
@@ -975,7 +975,7 @@ __mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
 *
 * This searches for pcie_device based on wwid, then return pcie_device object.
 */
struct _pcie_device *
static struct _pcie_device *
mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
{
	struct _pcie_device *pcie_device;
@@ -989,7 +989,7 @@ mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
}


struct _pcie_device *
static struct _pcie_device *
__mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id,
	int channel)
{
@@ -1012,34 +1012,7 @@ __mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id,
	return pcie_device;
}


/**
 * mpt3sas_get_pdev_by_idchannel - pcie device search
 * @ioc: per adapter object
 * @id: Target ID
 * @channel: Channel ID
 *
 * Context: This function will acquire ioc->pcie_device_lock and will release
 * before returning the pcie_device object.
 *
 * This searches for pcie_device based on id and channel, then return
 * pcie_device object.
 */
struct _pcie_device *
mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id, int channel)
{
	struct _pcie_device *pcie_device;
	unsigned long flags;

	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
	pcie_device = __mpt3sas_get_pdev_by_idchannel(ioc, id, channel);
	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);

	return pcie_device;
}


struct _pcie_device *
static struct _pcie_device *
__mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
{
	struct _pcie_device *pcie_device;