Loading drivers/net/ethernet/qualcomm/rmnet/rmnet_config.h +0 −1 Original line number Diff line number Diff line Loading @@ -34,7 +34,6 @@ struct rmnet_endpoint { */ struct rmnet_port { struct net_device *dev; struct rmnet_endpoint local_ep; struct rmnet_endpoint muxed_ep[RMNET_MAX_LOGICAL_EP]; u32 ingress_data_format; u32 egress_data_format; Loading drivers/net/ethernet/qualcomm/rmnet/rmnet_handlers.c +9 −68 Original line number Diff line number Diff line Loading @@ -44,27 +44,8 @@ static void rmnet_set_skb_proto(struct sk_buff *skb) /* Generic handler */ static rx_handler_result_t rmnet_bridge_handler(struct sk_buff *skb, struct rmnet_endpoint *ep) rmnet_deliver_skb(struct sk_buff *skb) { if (!ep->egress_dev) kfree_skb(skb); else rmnet_egress_handler(skb, ep); return RX_HANDLER_CONSUMED; } static rx_handler_result_t rmnet_deliver_skb(struct sk_buff *skb, struct rmnet_endpoint *ep) { switch (ep->rmnet_mode) { case RMNET_EPMODE_NONE: return RX_HANDLER_PASS; case RMNET_EPMODE_BRIDGE: return rmnet_bridge_handler(skb, ep); case RMNET_EPMODE_VND: skb_reset_transport_header(skb); skb_reset_network_header(skb); rmnet_vnd_rx_fixup(skb, skb->dev); Loading @@ -73,25 +54,6 @@ rmnet_deliver_skb(struct sk_buff *skb, struct rmnet_endpoint *ep) skb_set_mac_header(skb, 0); netif_receive_skb(skb); return RX_HANDLER_CONSUMED; default: kfree_skb(skb); return RX_HANDLER_CONSUMED; } } static rx_handler_result_t rmnet_ingress_deliver_packet(struct sk_buff *skb, struct rmnet_port *port) { if (!port) { kfree_skb(skb); return RX_HANDLER_CONSUMED; } skb->dev = port->local_ep.egress_dev; return rmnet_deliver_skb(skb, &port->local_ep); } /* MAP handler */ Loading Loading @@ -130,7 +92,7 @@ __rmnet_map_ingress_handler(struct sk_buff *skb, skb_pull(skb, sizeof(struct rmnet_map_header)); skb_trim(skb, len); rmnet_set_skb_proto(skb); return rmnet_deliver_skb(skb, ep); return rmnet_deliver_skb(skb); } static rx_handler_result_t Loading Loading @@ -204,29 +166,8 @@ rx_handler_result_t rmnet_rx_handler(struct sk_buff **pskb) dev = skb->dev; port = rmnet_get_port(dev); if (port->ingress_data_format & RMNET_INGRESS_FORMAT_MAP) { if (port->ingress_data_format & RMNET_INGRESS_FORMAT_MAP) rc = rmnet_map_ingress_handler(skb, port); } else { switch (ntohs(skb->protocol)) { case ETH_P_MAP: if (port->local_ep.rmnet_mode == RMNET_EPMODE_BRIDGE) { rc = rmnet_ingress_deliver_packet(skb, port); } else { kfree_skb(skb); rc = RX_HANDLER_CONSUMED; } break; case ETH_P_IP: case ETH_P_IPV6: rc = rmnet_ingress_deliver_packet(skb, port); break; default: rc = RX_HANDLER_PASS; } } return rc; } Loading drivers/net/ethernet/qualcomm/rmnet/rmnet_private.h +0 −8 Original line number Diff line number Diff line Loading @@ -19,23 +19,15 @@ #define RMNET_TX_QUEUE_LEN 1000 /* Constants */ #define RMNET_EGRESS_FORMAT__RESERVED__ BIT(0) #define RMNET_EGRESS_FORMAT_MAP BIT(1) #define RMNET_EGRESS_FORMAT_AGGREGATION BIT(2) #define RMNET_EGRESS_FORMAT_MUXING BIT(3) #define RMNET_EGRESS_FORMAT_MAP_CKSUMV3 BIT(4) #define RMNET_EGRESS_FORMAT_MAP_CKSUMV4 BIT(5) #define RMNET_INGRESS_FIX_ETHERNET BIT(0) #define RMNET_INGRESS_FORMAT_MAP BIT(1) #define RMNET_INGRESS_FORMAT_DEAGGREGATION BIT(2) #define RMNET_INGRESS_FORMAT_DEMUXING BIT(3) #define RMNET_INGRESS_FORMAT_MAP_COMMANDS BIT(4) #define RMNET_INGRESS_FORMAT_MAP_CKSUMV3 BIT(5) #define RMNET_INGRESS_FORMAT_MAP_CKSUMV4 BIT(6) /* Pass the frame up the stack with no modifications to skb->dev */ #define RMNET_EPMODE_NONE (0) /* Replace skb->dev to a virtual rmnet device and pass up the stack */ #define RMNET_EPMODE_VND (1) /* Pass the frame directly to another device with dev_queue_xmit() */ Loading Loading
drivers/net/ethernet/qualcomm/rmnet/rmnet_config.h +0 −1 Original line number Diff line number Diff line Loading @@ -34,7 +34,6 @@ struct rmnet_endpoint { */ struct rmnet_port { struct net_device *dev; struct rmnet_endpoint local_ep; struct rmnet_endpoint muxed_ep[RMNET_MAX_LOGICAL_EP]; u32 ingress_data_format; u32 egress_data_format; Loading
drivers/net/ethernet/qualcomm/rmnet/rmnet_handlers.c +9 −68 Original line number Diff line number Diff line Loading @@ -44,27 +44,8 @@ static void rmnet_set_skb_proto(struct sk_buff *skb) /* Generic handler */ static rx_handler_result_t rmnet_bridge_handler(struct sk_buff *skb, struct rmnet_endpoint *ep) rmnet_deliver_skb(struct sk_buff *skb) { if (!ep->egress_dev) kfree_skb(skb); else rmnet_egress_handler(skb, ep); return RX_HANDLER_CONSUMED; } static rx_handler_result_t rmnet_deliver_skb(struct sk_buff *skb, struct rmnet_endpoint *ep) { switch (ep->rmnet_mode) { case RMNET_EPMODE_NONE: return RX_HANDLER_PASS; case RMNET_EPMODE_BRIDGE: return rmnet_bridge_handler(skb, ep); case RMNET_EPMODE_VND: skb_reset_transport_header(skb); skb_reset_network_header(skb); rmnet_vnd_rx_fixup(skb, skb->dev); Loading @@ -73,25 +54,6 @@ rmnet_deliver_skb(struct sk_buff *skb, struct rmnet_endpoint *ep) skb_set_mac_header(skb, 0); netif_receive_skb(skb); return RX_HANDLER_CONSUMED; default: kfree_skb(skb); return RX_HANDLER_CONSUMED; } } static rx_handler_result_t rmnet_ingress_deliver_packet(struct sk_buff *skb, struct rmnet_port *port) { if (!port) { kfree_skb(skb); return RX_HANDLER_CONSUMED; } skb->dev = port->local_ep.egress_dev; return rmnet_deliver_skb(skb, &port->local_ep); } /* MAP handler */ Loading Loading @@ -130,7 +92,7 @@ __rmnet_map_ingress_handler(struct sk_buff *skb, skb_pull(skb, sizeof(struct rmnet_map_header)); skb_trim(skb, len); rmnet_set_skb_proto(skb); return rmnet_deliver_skb(skb, ep); return rmnet_deliver_skb(skb); } static rx_handler_result_t Loading Loading @@ -204,29 +166,8 @@ rx_handler_result_t rmnet_rx_handler(struct sk_buff **pskb) dev = skb->dev; port = rmnet_get_port(dev); if (port->ingress_data_format & RMNET_INGRESS_FORMAT_MAP) { if (port->ingress_data_format & RMNET_INGRESS_FORMAT_MAP) rc = rmnet_map_ingress_handler(skb, port); } else { switch (ntohs(skb->protocol)) { case ETH_P_MAP: if (port->local_ep.rmnet_mode == RMNET_EPMODE_BRIDGE) { rc = rmnet_ingress_deliver_packet(skb, port); } else { kfree_skb(skb); rc = RX_HANDLER_CONSUMED; } break; case ETH_P_IP: case ETH_P_IPV6: rc = rmnet_ingress_deliver_packet(skb, port); break; default: rc = RX_HANDLER_PASS; } } return rc; } Loading
drivers/net/ethernet/qualcomm/rmnet/rmnet_private.h +0 −8 Original line number Diff line number Diff line Loading @@ -19,23 +19,15 @@ #define RMNET_TX_QUEUE_LEN 1000 /* Constants */ #define RMNET_EGRESS_FORMAT__RESERVED__ BIT(0) #define RMNET_EGRESS_FORMAT_MAP BIT(1) #define RMNET_EGRESS_FORMAT_AGGREGATION BIT(2) #define RMNET_EGRESS_FORMAT_MUXING BIT(3) #define RMNET_EGRESS_FORMAT_MAP_CKSUMV3 BIT(4) #define RMNET_EGRESS_FORMAT_MAP_CKSUMV4 BIT(5) #define RMNET_INGRESS_FIX_ETHERNET BIT(0) #define RMNET_INGRESS_FORMAT_MAP BIT(1) #define RMNET_INGRESS_FORMAT_DEAGGREGATION BIT(2) #define RMNET_INGRESS_FORMAT_DEMUXING BIT(3) #define RMNET_INGRESS_FORMAT_MAP_COMMANDS BIT(4) #define RMNET_INGRESS_FORMAT_MAP_CKSUMV3 BIT(5) #define RMNET_INGRESS_FORMAT_MAP_CKSUMV4 BIT(6) /* Pass the frame up the stack with no modifications to skb->dev */ #define RMNET_EPMODE_NONE (0) /* Replace skb->dev to a virtual rmnet device and pass up the stack */ #define RMNET_EPMODE_VND (1) /* Pass the frame directly to another device with dev_queue_xmit() */ Loading