Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit f0dfebaf authored by Sreenivasa Honnur's avatar Sreenivasa Honnur Committed by David S. Miller
Browse files

vxge: Removed unused functions.



- Removed the wrr_rebalance function

- This feature is not supported by the ASIC, hence removing the related code.

Signed-off-by: default avatarSreenivasa Honnur <sreenivasa.honnur@neterion.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent a4a987d8
Loading
Loading
Loading
Loading
+0 −204
Original line number Diff line number Diff line
@@ -460,209 +460,6 @@ __vxge_hw_verify_pci_e_info(struct __vxge_hw_device *hldev)
	return VXGE_HW_OK;
}

/*
 * vxge_hw_wrr_rebalance - Rebalance the RX_WRR and KDFC_WRR calandars.
 * Rebalance the RX_WRR and KDFC_WRR calandars.
 */
static enum
vxge_hw_status vxge_hw_wrr_rebalance(struct __vxge_hw_device *hldev)
{
	u64 val64;
	u32 wrr_states[VXGE_HW_WEIGHTED_RR_SERVICE_STATES];
	u32 i, j, how_often = 1;
	enum vxge_hw_status status = VXGE_HW_OK;

	status = __vxge_hw_device_is_privilaged(hldev->host_type,
			hldev->func_id);
	if (status != VXGE_HW_OK)
		goto exit;

	/* Reset the priorities assigned to the WRR arbitration
	phases for the receive traffic */
	for (i = 0; i < VXGE_HW_WRR_RING_COUNT; i++)
		writeq(0, ((&hldev->mrpcim_reg->rx_w_round_robin_0) + i));

	/* Reset the transmit FIFO servicing calendar for FIFOs */
	for (i = 0; i < VXGE_HW_WRR_FIFO_COUNT; i++) {
		writeq(0, ((&hldev->mrpcim_reg->kdfc_w_round_robin_0) + i));
		writeq(0, ((&hldev->mrpcim_reg->kdfc_w_round_robin_20) + i));
	}

	/* Assign WRR priority  0 for all FIFOs */
	for (i = 1; i < VXGE_HW_MAX_VIRTUAL_PATHS; i++) {
		writeq(VXGE_HW_KDFC_FIFO_0_CTRL_WRR_NUMBER(0),
				((&hldev->mrpcim_reg->kdfc_fifo_0_ctrl)  + i));

		writeq(VXGE_HW_KDFC_FIFO_17_CTRL_WRR_NUMBER(0),
			((&hldev->mrpcim_reg->kdfc_fifo_17_ctrl) + i));
	}

	/* Reset to service non-offload doorbells */
	writeq(0, &hldev->mrpcim_reg->kdfc_entry_type_sel_0);
	writeq(0, &hldev->mrpcim_reg->kdfc_entry_type_sel_1);

	/* Set priority 0 to all receive queues */
	writeq(0, &hldev->mrpcim_reg->rx_queue_priority_0);
	writeq(0, &hldev->mrpcim_reg->rx_queue_priority_1);
	writeq(0, &hldev->mrpcim_reg->rx_queue_priority_2);

	/* Initialize all the slots as unused */
	for (i = 0; i < VXGE_HW_WEIGHTED_RR_SERVICE_STATES; i++)
		wrr_states[i] = -1;

	/* Prepare the Fifo service states */
	for (i = 0; i < VXGE_HW_MAX_VIRTUAL_PATHS; i++) {

		if (!hldev->config.vp_config[i].min_bandwidth)
			continue;

		how_often = VXGE_HW_VPATH_BANDWIDTH_MAX /
				hldev->config.vp_config[i].min_bandwidth;
		if (how_often) {

			for (j = 0; j < VXGE_HW_WRR_FIFO_SERVICE_STATES;) {
				if (wrr_states[j] == -1) {
					wrr_states[j] = i;
					/* Make sure each fifo is serviced
					 * atleast once */
					if (i == j)
						j += VXGE_HW_MAX_VIRTUAL_PATHS;
					else
						j += how_often;
				} else
					j++;
			}
		}
	}

	/* Fill the unused slots with 0 */
	for (j = 0; j < VXGE_HW_WEIGHTED_RR_SERVICE_STATES; j++) {
		if (wrr_states[j] == -1)
			wrr_states[j] = 0;
	}

	/* Assign WRR priority number for FIFOs */
	for (i = 0; i < VXGE_HW_MAX_VIRTUAL_PATHS; i++) {
		writeq(VXGE_HW_KDFC_FIFO_0_CTRL_WRR_NUMBER(i),
				((&hldev->mrpcim_reg->kdfc_fifo_0_ctrl) + i));

		writeq(VXGE_HW_KDFC_FIFO_17_CTRL_WRR_NUMBER(i),
			((&hldev->mrpcim_reg->kdfc_fifo_17_ctrl) + i));
	}

	/* Modify the servicing algorithm applied to the 3 types of doorbells.
	i.e, none-offload, message and offload */
	writeq(VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_0(0) |
				VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_1(0) |
				VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_2(0) |
				VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_3(0) |
				VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_4(1) |
				VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_5(0) |
				VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_6(0) |
				VXGE_HW_KDFC_ENTRY_TYPE_SEL_0_NUMBER_7(0),
				&hldev->mrpcim_reg->kdfc_entry_type_sel_0);

	writeq(VXGE_HW_KDFC_ENTRY_TYPE_SEL_1_NUMBER_8(1),
				&hldev->mrpcim_reg->kdfc_entry_type_sel_1);

	for (i = 0, j = 0; i < VXGE_HW_WRR_FIFO_COUNT; i++) {

		val64 = VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_0(wrr_states[j++]);
		val64 |= VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_1(wrr_states[j++]);
		val64 |= VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_2(wrr_states[j++]);
		val64 |= VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_3(wrr_states[j++]);
		val64 |= VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_4(wrr_states[j++]);
		val64 |= VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_5(wrr_states[j++]);
		val64 |= VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_6(wrr_states[j++]);
		val64 |= VXGE_HW_KDFC_W_ROUND_ROBIN_0_NUMBER_7(wrr_states[j++]);

		writeq(val64, (&hldev->mrpcim_reg->kdfc_w_round_robin_0 + i));
		writeq(val64, (&hldev->mrpcim_reg->kdfc_w_round_robin_20 + i));
	}

	/* Set up the priorities assigned to receive queues */
	writeq(VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_0(0) |
			VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_1(1) |
			VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_2(2) |
			VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_3(3) |
			VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_4(4) |
			VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_5(5) |
			VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_6(6) |
			VXGE_HW_RX_QUEUE_PRIORITY_0_RX_Q_NUMBER_7(7),
			&hldev->mrpcim_reg->rx_queue_priority_0);

	writeq(VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_8(8) |
			VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_9(9) |
			VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_10(10) |
			VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_11(11) |
			VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_12(12) |
			VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_13(13) |
			VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_14(14) |
			VXGE_HW_RX_QUEUE_PRIORITY_1_RX_Q_NUMBER_15(15),
			&hldev->mrpcim_reg->rx_queue_priority_1);

	writeq(VXGE_HW_RX_QUEUE_PRIORITY_2_RX_Q_NUMBER_16(16),
				&hldev->mrpcim_reg->rx_queue_priority_2);

	/* Initialize all the slots as unused */
	for (i = 0; i < VXGE_HW_WEIGHTED_RR_SERVICE_STATES; i++)
		wrr_states[i] = -1;

	/* Prepare the Ring service states */
	for (i = 0; i < VXGE_HW_MAX_VIRTUAL_PATHS; i++) {

		if (!hldev->config.vp_config[i].min_bandwidth)
			continue;

		how_often = VXGE_HW_VPATH_BANDWIDTH_MAX /
				hldev->config.vp_config[i].min_bandwidth;

		if (how_often) {
			for (j = 0; j < VXGE_HW_WRR_RING_SERVICE_STATES;) {
				if (wrr_states[j] == -1) {
					wrr_states[j] = i;
					/* Make sure each ring is
					 * serviced atleast once */
					if (i == j)
						j += VXGE_HW_MAX_VIRTUAL_PATHS;
					else
						j += how_often;
				} else
					j++;
			}
		}
	}

	/* Fill the unused slots with 0 */
	for (j = 0; j < VXGE_HW_WEIGHTED_RR_SERVICE_STATES; j++) {
		if (wrr_states[j] == -1)
			wrr_states[j] = 0;
	}

	for (i = 0, j = 0; i < VXGE_HW_WRR_RING_COUNT; i++) {
		val64 =  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_0(
				wrr_states[j++]);
		val64 |=  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_1(
				wrr_states[j++]);
		val64 |=  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_2(
				wrr_states[j++]);
		val64 |=  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_3(
				wrr_states[j++]);
		val64 |=  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_4(
				wrr_states[j++]);
		val64 |=  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_5(
				wrr_states[j++]);
		val64 |=  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_6(
				wrr_states[j++]);
		val64 |=  VXGE_HW_RX_W_ROUND_ROBIN_0_RX_W_PRIORITY_SS_7(
				wrr_states[j++]);

		writeq(val64, ((&hldev->mrpcim_reg->rx_w_round_robin_0) + i));
	}
exit:
	return status;
}

/*
 * __vxge_hw_device_initialize
 * Initialize Titan-V hardware.
@@ -679,7 +476,6 @@ enum vxge_hw_status __vxge_hw_device_initialize(struct __vxge_hw_device *hldev)
			goto exit;
	}

	vxge_hw_wrr_rebalance(hldev);
exit:
	return status;
}