/* * IUCV network driver * * Copyright IBM Corp. 2001, 2009 * * Author(s): * Original netiucv driver: * Fritz Elfert (elfert@de.ibm.com, felfert@millenux.com) * Sysfs integration and all bugs therein: * Cornelia Huck (cornelia.huck@de.ibm.com) * PM functions: * Ursula Braun (ursula.braun@de.ibm.com) * * Documentation used: * the source of the original IUCV driver by: * Stefan Hegewald <hegewald@de.ibm.com> * Hartmut Penner <hpenner@de.ibm.com> * Denis Joseph Barrow (djbarrow@de.ibm.com,barrow_dj@yahoo.com) * Martin Schwidefsky (schwidefsky@de.ibm.com) * Alan Altmark (Alan_Altmark@us.ibm.com) Sept. 2000 * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2, or (at your option) * any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. * */ #define KMSG_COMPONENT "netiucv" #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt #undef DEBUG #include <linux/module.h> #include <linux/init.h> #include <linux/kernel.h> #include <linux/slab.h> #include <linux/errno.h> #include <linux/types.h> #include <linux/interrupt.h> #include <linux/timer.h> #include <linux/bitops.h> #include <linux/signal.h> #include <linux/string.h> #include <linux/device.h> #include <linux/ip.h> #include <linux/if_arp.h> #include <linux/tcp.h> #include <linux/skbuff.h> #include <linux/ctype.h> #include <net/dst.h> #include <asm/io.h> #include <asm/uaccess.h> #include <asm/ebcdic.h> #include <net/iucv/iucv.h> #include "fsm.h" MODULE_AUTHOR ("(C) 2001 IBM Corporation by Fritz Elfert (felfert@millenux.com)"); MODULE_DESCRIPTION ("Linux for S/390 IUCV network driver"); /** * Debug Facility stuff */ #define IUCV_DBF_SETUP_NAME "iucv_setup" #define IUCV_DBF_SETUP_LEN 64 #define IUCV_DBF_SETUP_PAGES 2 #define IUCV_DBF_SETUP_NR_AREAS 1 #define IUCV_DBF_SETUP_LEVEL 3 #define IUCV_DBF_DATA_NAME "iucv_data" #define IUCV_DBF_DATA_LEN 128 #define IUCV_DBF_DATA_PAGES 2 #define IUCV_DBF_DATA_NR_AREAS 1 #define IUCV_DBF_DATA_LEVEL 2 #define IUCV_DBF_TRACE_NAME "iucv_trace" #define IUCV_DBF_TRACE_LEN 16 #define IUCV_DBF_TRACE_PAGES 4 #define IUCV_DBF_TRACE_NR_AREAS 1 #define IUCV_DBF_TRACE_LEVEL 3 #define IUCV_DBF_TEXT(name,level,text) \ do { \ debug_text_event(iucv_dbf_##name,level,text); \ } while (0) #define IUCV_DBF_HEX(name,level,addr,len) \ do { \ debug_event(iucv_dbf_##name,level,(void*)(addr),len); \ } while (0) DECLARE_PER_CPU(char[256], iucv_dbf_txt_buf); /* Allow to sort out low debug levels early to avoid wasted sprints */ static inline int iucv_dbf_passes(debug_info_t *dbf_grp, int level) { return (level <= dbf_grp->level); } #define IUCV_DBF_TEXT_(name, level, text...) \ do { \ if (iucv_dbf_passes(iucv_dbf_##name, level)) { \ char* __buf = get_cpu_var(iucv_dbf_txt_buf); \ sprintf(__buf, text); \ debug_text_event(iucv_dbf_##name, level, __buf); \ put_cpu_var(iucv_dbf_txt_buf); \ } \ } while (0) #define IUCV_DBF_SPRINTF(name,level,text...) \ do { \ debug_sprintf_event(iucv_dbf_trace, level, ##text ); \ debug_sprintf_event(iucv_dbf_trace, level, text ); \ } while (0) /** * some more debug stuff */ #define IUCV_HEXDUMP16(importance,header,ptr) \ PRINT_##importance(header "%02x %02x %02x %02x %02x %02x %02x %02x " \ "%02x %02x %02x %02x %02x %02x %02x %02x\n", \ *(((char*)ptr)),*(((char*)ptr)+1),*(((char*)ptr)+2), \ *(((char*)ptr)+3),*(((char*)ptr)+4),*(((char*)ptr)+5), \ *(((char*)ptr)+6),*(((char*)ptr)+7),*(((char*)ptr)+8), \ *(((char*)ptr)+9),*(((char*)ptr)+10),*(((char*)ptr)+11), \ *(((char*)ptr)+12),*(((char*)ptr)+13), \ *(((char*)ptr)+14),*(((char*)ptr)+15)); \ PRINT_##importance(header "%02x %02x %02x %02x %02x %02x %02x %02x " \ "%02x %02x %02x %02x %02x %02x %02x %02x\n", \ *(((char*)ptr)+16),*(((char*)ptr)+17), \ *(((char*)ptr)+18),*(((char*)ptr)+19), \ *(((char*)ptr)+20),*(((char*)ptr)+21), \ *(((char*)ptr)+22),*(((char*)ptr)+23), \ *(((char*)ptr)+24),*(((char*)ptr)+25), \ *(((char*)ptr)+26),*(((char*)ptr)+27), \ *(((char*)ptr)+28),*(((char*)ptr)+29), \ *(((char*)ptr)+30),*(((char*)ptr)+31)); #define PRINTK_HEADER " iucv: " /* for debugging */ /* dummy device to make sure netiucv_pm functions are called */ static struct device *netiucv_dev; static int netiucv_pm_prepare(struct device *); static void netiucv_pm_complete(struct device *); static int netiucv_pm_freeze(struct device *); static int netiucv_pm_restore_thaw(struct device *); static const struct dev_pm_ops netiucv_pm_ops = { .prepare = netiucv_pm_prepare, .complete = netiucv_pm_complete, .freeze = netiucv_pm_freeze, .thaw = netiucv_pm_restore_thaw, .restore = netiucv_pm_restore_thaw, }; static struct device_driver netiucv_driver = { .owner = THIS_MODULE, .name = "netiucv", .bus = &iucv_bus, .pm = &netiucv_pm_ops, }; static int netiucv_callback_connreq(struct iucv_path *, u8 ipvmid[8], u8 ipuser[16]); static void netiucv_callback_connack(struct iucv_path *, u8 ipuser[16]); static void netiucv_callback_connrej(struct iucv_path *, u8 ipuser[16]); static void netiucv_callback_connsusp(struct iucv_path *, u8 ipuser[16]); static void netiucv_callback_connres(struct iucv_path *, u8 ipuser[16]); static void netiucv_callback_rx(struct iucv_path *, struct iucv_message *); static void netiucv_callback_txdone(struct iucv_path *, struct iucv_message *); static struct iucv_handler netiucv_handler = { .path_pending = netiucv_callback_connreq, .path_complete = netiucv_callback_connack, .path_severed = netiucv_callback_connrej, .path_quiesced = netiucv_callback_connsusp, .path_resumed = netiucv_callback_connres, .message_pending = netiucv_callback_rx, .message_complete = netiucv_callback_txdone }; /** * Per connection profiling data */ struct connection_profile { unsigned long maxmulti; unsigned long maxcqueue; unsigned long doios_single; unsigned long doios_multi; unsigned long txlen; unsigned long tx_time; struct timespec send_stamp; unsigned long tx_pending; unsigned long tx_max_pending; }; /** * Representation of one iucv connection */ struct iucv_connection { struct list_head list; struct iucv_path *path; struct sk_buff *rx_buff; struct sk_buff *tx_buff; struct sk_buff_head collect_queue; struct sk_buff_head commit_queue; spinlock_t collect_lock; int collect_len; int max_buffsize; fsm_timer timer; fsm_instance *fsm; struct net_device *netdev; struct connection_profile prof; char userid[9]; char userdata[17]; }; /** * Linked list of all connection structs. */ static LIST_HEAD(iucv_connection_list); static DEFINE_RWLOCK(iucv_connection_rwlock); /** * Representation of event-data for the * connection state machine. */ struct iucv_event { struct iucv_connection *conn; void *data; }; /** * Private part of the network device structure */ struct netiucv_priv { struct net_device_stats stats; unsigned long tbusy; fsm_instance *fsm; struct iucv_connection *conn; struct device *dev; int pm_state; }; /** * Link level header for a packet. */ struct ll_header { u16 next; }; #define NETIUCV_HDRLEN (sizeof(struct ll_header)) #define NETIUCV_BUFSIZE_MAX 65537 #define NETIUCV_BUFSIZE_DEFAULT NETIUCV_BUFSIZE_MAX #define NETIUCV_MTU_MAX (NETIUCV_BUFSIZE_MAX - NETIUCV_HDRLEN) #define NETIUCV_MTU_DEFAULT 9216 #define NETIUCV_QUEUELEN_DEFAULT 50 #define NETIUCV_TIMEOUT_5SEC 5000 /** * Compatibility macros for busy handling * of network devices. */ static inline void netiucv_clear_busy(struct net_device *dev) { struct netiucv_priv *priv = netdev_priv(dev); clear_bit(0, &priv->tbusy); netif_wake_queue(dev); } static inline int netiucv_test_and_set_busy(struct net_device *dev) { struct netiucv_priv *priv = netdev_priv(dev); netif_stop_queue(dev); return test_and_set_bit(0, &priv->tbusy); } static u8 iucvMagic_ascii[16] = { 0x30, 0x20, 0x20, 0x20, 0x20, 0x20, 0x20, 0x20, 0x30, 0x20, 0x20, 0x20, 0x20, 0x20, 0x20, 0x20 }; static u8 iucvMagic_ebcdic[16] = { 0xF0, 0x40, 0x40, 0x40, 0x40, 0x40, 0x40, 0x40, 0xF0, 0x40, 0x40, 0x40, 0x40, 0x40, 0x40, 0x40 }; /** * Convert an iucv userId to its printable * form (strip whitespace at end). * * @param An iucv userId * * @returns The printable string (static data!!) */ static char *netiucv_printname(char *name, int len) { static char tmp[17]; char *p = tmp; memcpy(tmp, name, len); tmp[len] = '\0'; while (*p && ((p - tmp) < len) && (!isspace(*p))) p++; *p = '\0'; return tmp; } static char *netiucv_printuser(struct iucv_connection *conn) { static char tmp_uid[9]; static char tmp_udat[17]; static char buf[100]; if (memcmp(conn->userdata, iucvMagic_ebcdic, 16)) { tmp_uid[8] = '\0'; tmp_udat[16] = '\0'; memcpy(tmp_uid, conn->userid, 8); memcpy(tmp_uid, netiucv_printname(tmp_uid, 8), 8); memcpy(tmp_udat, conn->userdata, 16); EBCASC(tmp_udat, 16); memcpy(tmp_udat, netiucv_printname(tmp_udat, 16), 16); sprintf(buf, "%s.%s", tmp_uid, tmp_udat); return buf; } else return netiucv_printname(conn->userid, 8); } /** * States of the interface statemachine. */ enum dev_states { DEV_STATE_STOPPED, DEV_STATE_STARTWAIT, DEV_STATE_STOPWAIT, DEV_STATE_RUNNING, /** * MUST be always the last element!! */ NR_DEV_STATES }; static const char *dev_state_names[] = { "Stopped", "StartWait", "StopWait", "Running", }; /** * Events of the interface statemachine. */ enum dev_events { DEV_EVENT_START, DEV_EVENT_STOP, DEV_EVENT_CONUP, DEV_EVENT_CONDOWN, /** * MUST be always the last element!! */ NR_DEV_EVENTS }; static const char *dev_event_names[] = { "Start", "Stop", "Connection up", "Connection down", }; /** * Events of the connection statemachine */ enum conn_events { /** * Events, representing callbacks from * lowlevel iucv layer) */ CONN_EVENT_CONN_REQ, CONN_EVENT_CONN_ACK, CONN_EVENT_CONN_REJ, CONN_EVENT_CONN_SUS, CONN_EVENT_CONN_RES, CONN_EVENT_RX, CONN_EVENT_TXDONE, /** * Events, representing errors return codes from * calls to lowlevel iucv layer */ /** * Event, representing timer expiry. */ CONN_EVENT_TIMER, /** * Events, representing commands from upper levels. */ CONN_EVENT_START, CONN_EVENT_STOP, /** * MUST be always the last element!! */ NR_CONN_EVENTS, }; static const char *conn_event_names[] = { "Remote connection request", "Remote connection acknowledge", "Remote connection reject", "Connection suspended", "Connection resumed", "Data received", "Data sent", "Timer", "Start", "Stop", }; /** * States of the connection statemachine. */ enum conn_states { /** * Connection not assigned to any device, * initial state, invalid */ CONN_STATE_INVALID, /** * Userid assigned but not operating */ CONN_STATE_STOPPED, /** * Connection registered, * no connection request sent yet, * no connection request received */ CONN_STATE_STARTWAIT, /** * Connection registered and connection request sent, * no acknowledge and no connection request received yet. */ CONN_STATE_SETUPWAIT, /** * Connection up and running idle */ CONN_STATE_IDLE, /** * Data sent, awaiting CONN_EVENT_TXDONE */ CONN_STATE_TX, /** * Error during registration. */ CONN_STATE_REGERR, /** * Error during registration. */ CONN_STATE_CONNERR, /** * MUST be always the last element!! */ NR_CONN_STATES, }; static const char *conn_state_names[] = { "Invalid", "Stopped", "StartWait", "SetupWait", "Idle", "TX", "Terminating", "Registration error", "Connect error", }; /** * Debug Facility Stuff */ static debug_info_t *iucv_dbf_setup = NULL; static debug_info_t *iucv_dbf_data = NULL; static debug_info_t *iucv_dbf_trace = NULL; DEFINE_PER_CPU(char[256], iucv_dbf_txt_buf); static void iucv_unregister_dbf_views(void) { if (iucv_dbf_setup) debug_unregister(iucv_dbf_setup); if (iucv_dbf_data) debug_unregister(iucv_dbf_data); if (iucv_dbf_trace) debug_unregister(iucv_dbf_trace); } static int iucv_register_dbf_views(void) { iucv_dbf_setup = debug_register(IUCV_DBF_SETUP_NAME, IUCV_DBF_SETUP_PAGES, IUCV_DBF_SETUP_NR_AREAS, IUCV_DBF_SETUP_LEN); iucv_dbf_data = debug_register(IUCV_DBF_DATA_NAME, IUCV_DBF_DATA_PAGES, IUCV_DBF_DATA_NR_AREAS, IUCV_DBF_DATA_LEN); iucv_dbf_trace = debug_register(IUCV_DBF_TRACE_NAME, IUCV_DBF_TRACE_PAGES, IUCV_DBF_TRACE_NR_AREAS, IUCV_DBF_TRACE_LEN); if ((iucv_dbf_setup == NULL) || (iucv_dbf_data == NULL) || (iucv_dbf_trace == NULL)) { iucv_unregister_dbf_views(); return -ENOMEM; } debug_register_view(iucv_dbf_setup, &debug_hex_ascii_view); debug_set_level(iucv_dbf_setup, IUCV_DBF_SETUP_LEVEL); debug_register_view(iucv_dbf_data, &debug_hex_ascii_view); debug_set_level(iucv_dbf_data, IUCV_DBF_DATA_LEVEL); debug_register_view(iucv_dbf_trace, &debug_hex_ascii_view); debug_set_level(iucv_dbf_trace, IUCV_DBF_TRACE_LEVEL); return 0; } /* * Callback-wrappers, called from lowlevel iucv layer. */ static void netiucv_callback_rx(struct iucv_path *path, struct iucv_message *msg) { struct iucv_connection *conn = path->private; struct iucv_event ev; ev.conn = conn; ev.data = msg; fsm_event(conn->fsm, CONN_EVENT_RX, &ev); } static void netiucv_callback_txdone(struct iucv_path *path, struct iucv_message *msg) { struct iucv_connection *conn = path->private; struct iucv_event ev; ev.conn = conn; ev.data = msg; fsm_event(conn->fsm, CONN_EVENT_TXDONE, &ev); } static void netiucv_callback_connack(struct iucv_path *path, u8 ipuser[16]) { struct iucv_connection *conn = path->private; fsm_event(conn->fsm, CONN_EVENT_CONN_ACK, conn); } static int netiucv_callback_connreq(struct iucv_path *path, u8 ipvmid[8], u8 ipuser[16]) { struct iucv_connection *conn = path->private; struct iucv_event ev; static char tmp_user[9]; static char tmp_udat[17]; int rc; rc = -EINVAL; memcpy(tmp_user, netiucv_printname(ipvmid, 8), 8); memcpy(tmp_udat, ipuser, 16); EBCASC(tmp_udat, 16); read_lock_bh(&iucv_connection_rwlock); list_for_each_entry(conn, &iucv_connection_list, list) { if (strncmp(ipvmid, conn->userid, 8) || strncmp(ipuser, conn->userdata, 16)) continue; /* Found a matching connection for this path. */ conn->path = path; ev.conn = conn; ev.data = path; fsm_event(conn->fsm, CONN_EVENT_CONN_REQ, &ev); rc = 0; } IUCV_DBF_TEXT_(setup, 2, "Connection requested for %s.%s\n", tmp_user, netiucv_printname(tmp_udat, 16)); read_unlock_bh(&iucv_connection_rwlock); return rc; } static void netiucv_callback_connrej(struct iucv_path *path, u8 ipuser[16]) { struct iucv_connection *conn = path->private; fsm_event(conn->fsm, CONN_EVENT_CONN_REJ, conn); } static void netiucv_callback_connsusp(struct iucv_path *path, u8 ipuser[16]) { struct iucv_connection *conn = path->private; fsm_event(conn->fsm, CONN_EVENT_CONN_SUS, conn); } static void netiucv_callback_connres(struct iucv_path *path, u8 ipuser[16]) { struct iucv_connection *conn = path->private; fsm_event(conn->fsm, CONN_EVENT_CONN_RES, conn); } /** * NOP action for statemachines */ static void netiucv_action_nop(fsm_instance *fi, int event, void *arg) { } /* * Actions of the connection statemachine */ /** * netiucv_unpack_skb * @conn: The connection where this skb has been received. * @pskb: The received skb. * * Unpack a just received skb and hand it over to upper layers. * Helper function for conn_action_rx. */ static void netiucv_unpack_skb(struct iucv_connection *conn, struct sk_buff *pskb) { struct net_device *dev = conn->netdev; struct netiucv_priv *privptr = netdev_priv(dev); u16 offset = 0; skb_put(pskb, NETIUCV_HDRLEN); pskb->dev = dev; pskb->ip_summed = CHECKSUM_NONE; pskb->protocol = ntohs(ETH_P_IP); while (1) { struct sk_buff *skb; struct ll_header *header = (struct ll_header *) pskb->data; if (!header->next) break; skb_pull(pskb, NETIUCV_HDRLEN); header->next -= offset; offset += header->next; header->next -= NETIUCV_HDRLEN; if (skb_tailroom(pskb) < header->next) { IUCV_DBF_TEXT_(data, 2, "Illegal next field: %d > %d\n", header->next, skb_tailroom(pskb)); return; } skb_put(pskb, header->next); skb_reset_mac_header(pskb); skb = dev_alloc_skb(pskb->len); if (!skb) { IUCV_DBF_TEXT(data, 2, "Out of memory in netiucv_unpack_skb\n"); privptr->stats.rx_dropped++; return; } skb_copy_from_linear_data(pskb, skb_put(skb, pskb->len), pskb->len); skb_reset_mac_header(skb); skb->dev = pskb->dev; skb->protocol = pskb->protocol; pskb->ip_summed = CHECKSUM_UNNECESSARY; privptr->stats.rx_packets++; privptr->stats.rx_bytes += skb->len; /* * Since receiving is always initiated from a tasklet (in iucv.c), * we must use netif_rx_ni() instead of netif_rx() */ netif_rx_ni(skb); skb_pull(pskb, header->next); skb_put(pskb, NETIUCV_HDRLEN); } } static void conn_action_rx(fsm_instance *fi, int event, void *arg) { struct iucv_event *ev = arg; struct iucv_connection *conn = ev->conn; struct iucv_message *msg = ev->data; struct netiucv_priv *privptr = netdev_priv(conn->netdev); int rc; IUCV_DBF_TEXT(trace, 4, __func__); if (!conn->netdev) { iucv_message_reject(conn->path, msg); IUCV_DBF_TEXT(data, 2, "Received data for unlinked connection\n"); return; } if (msg->length > conn->max_buffsize) { iucv_message_reject(conn->path, msg); privptr->stats.rx_dropped++; IUCV_DBF_TEXT_(data, 2, "msglen %d > max_buffsize %d\n", msg->length, conn->max_buffsize); return; } conn->rx_buff->data = conn->rx_buff->head; skb_reset_tail_pointer(conn->rx_buff); conn->rx_buff->len = 0; rc = iucv_message_receive(conn->path, msg, 0, conn->rx_buff->data, msg->length, NULL); if (rc || msg->length < 5) { privptr->stats.rx_errors++; IUCV_DBF_TEXT_(data, 2, "rc %d from iucv_receive\n", rc); return; } netiucv_unpack_skb(conn, conn->rx_buff); } static void conn_action_txdone(fsm_instance *fi, int event, void *arg) { struct iucv_event *ev = arg; struct iucv_connection *conn = ev->conn; struct iucv_message *msg = ev->data; struct iucv_message txmsg; struct netiucv_priv *privptr = NULL; u32 single_flag = msg->tag; u32 txbytes = 0; u32 txpackets = 0; u32 stat_maxcq = 0; struct sk_buff *skb; unsigned long saveflags; struct ll_header header; int rc; IUCV_DBF_TEXT(trace, 4, __func__); if (conn && conn->netdev) privptr = netdev_priv(conn->netdev); conn->prof.tx_pending--; if (single_flag) { if ((skb = skb_dequeue(&conn->commit_queue))) { atomic_dec(&skb->users); if (privptr) { privptr->stats.tx_packets++; privptr->stats.tx_bytes += (skb->len - NETIUCV_HDRLEN - NETIUCV_HDRLEN); } dev_kfree_skb_any(skb); } } conn->tx_buff->data = conn->tx_buff->head; skb_reset_tail_pointer(conn->tx_buff); conn->tx_buff->len = 0; spin_lock_irqsave(&conn->collect_lock, saveflags); while ((skb = skb_dequeue(&conn->collect_queue))) { header.next = conn->tx_buff->len + skb->len + NETIUCV_HDRLEN; memcpy(skb_put(conn->tx_buff, NETIUCV_HDRLEN), &header, NETIUCV_HDRLEN); skb_copy_from_linear_data(skb, skb_put(conn->tx_buff, skb->len), skb->len); txbytes += skb->len; txpackets++; stat_maxcq++; atomic_dec(&skb->users); dev_kfree_skb_any(skb); } if (conn->collect_len > conn->prof.maxmulti) conn->prof.maxmulti = conn->collect_len; conn->collect_len = 0; spin_unlock_irqrestore(&conn->collect_lock, saveflags); if (conn->tx_buff->len == 0) { fsm_newstate(fi, CONN_STATE_IDLE); return; } header.next = 0; memcpy(skb_put(conn->tx_buff, NETIUCV_HDRLEN), &header, NETIUCV_HDRLEN); conn->prof.send_stamp = current_kernel_time(); txmsg.class = 0; txmsg.tag = 0; rc = iucv_message_send(conn->path, &txmsg, 0, 0, conn->tx_buff->data, conn->tx_buff->len); conn->prof.doios_multi++; conn->prof.txlen += conn->tx_buff->len; conn->prof.tx_pending++; if (conn->prof.tx_pending > conn->prof.tx_max_pending) conn->prof.tx_max_pending = conn->prof.tx_pending; if (rc) { conn->prof.tx_pending--; fsm_newstate(fi, CONN_STATE_IDLE); if (privptr) privptr->stats.tx_errors += txpackets; IUCV_DBF_TEXT_(data, 2, "rc %d from iucv_send\n", rc); } else { if (privptr) { privptr->stats.tx_packets += txpackets; privptr->stats.tx_bytes += txbytes; } if (stat_maxcq > conn->prof.maxcqueue) conn->prof.maxcqueue = stat_maxcq; } } static void conn_action_connaccept(fsm_instance *fi, int event, void *arg) { struct iucv_event *ev = arg; struct iucv_connection *conn = ev->conn; struct iucv_path *path = ev->data; struct net_device *netdev = conn->netdev; struct netiucv_priv *privptr = netdev_priv(netdev); int rc; IUCV_DBF_TEXT(trace, 3, __func__); conn->path = path; path->msglim = NETIUCV_QUEUELEN_DEFAULT; path->flags = 0; rc = iucv_path_accept(path, &netiucv_handler, conn->userdata , conn); if (rc) { IUCV_DBF_TEXT_(setup, 2, "rc %d from iucv_accept", rc); return; } fsm_newstate(fi, CONN_STATE_IDLE); netdev->tx_queue_len = conn->path->msglim; fsm_event(privptr->fsm, DEV_EVENT_CONUP, netdev); } static void conn_action_connreject(fsm_instance *fi, int event, void *arg) { struct iucv_event *ev = arg; struct iucv_path *path = ev->data; IUCV_DBF_TEXT(trace, 3, __func__); iucv_path_sever(path, NULL); } static void conn_action_connack(fsm_instance *fi, int event, void *arg) { struct iucv_connection *conn = arg; struct net_device *netdev = conn->netdev; struct netiucv_priv *privptr = netdev_priv(netdev); IUCV_DBF_TEXT(trace, 3, __func__); fsm_deltimer(&conn->timer); fsm_newstate(fi, CONN_STATE_IDLE); netdev->tx_queue_len = conn->path->msglim; fsm_event(privptr->fsm, DEV_EVENT_CONUP, netdev); } static void conn_action_conntimsev(fsm_instance *fi, int event, void *arg) { struct iucv_connection *conn = arg; IUCV_DBF_TEXT(trace, 3, __func__); fsm_deltimer(&conn->timer); iucv_path_sever(conn->path, conn->userdata); fsm_newstate(fi, CONN_STATE_STARTWAIT); } static void conn_action_connsever(fsm_instance *fi, int event, void *arg) { struct iucv_connection *conn = arg; struct net_device *netdev = conn->netdev; struct netiucv_priv *privptr = netdev_priv(netdev); IUCV_DBF_TEXT(trace, 3, __func__); fsm_deltimer(&conn->timer); iucv_path_sever(conn->path, conn->userdata); dev_info(privptr->dev, "The peer z/VM guest %s has closed the " "connection\n", netiucv_printuser(conn)); IUCV_DBF_TEXT(data, 2, "conn_action_connsever: Remote dropped connection\n"); fsm_newstate(fi, CONN_STATE_STARTWAIT); fsm_event(privptr->fsm, DEV_EVENT_CONDOWN, netdev); } static void conn_action_start(fsm_instance *fi, int event, void *arg) { struct iucv_connection *conn = arg; struct net_device *netdev = conn->netdev; struct netiucv_priv *privptr = netdev_priv(netdev); int rc; IUCV_DBF_TEXT(trace, 3, __func__); fsm_newstate(fi, CONN_STATE_STARTWAIT); /* * We must set the state before calling iucv_connect because the * callback handler could be called at any point after the connection * request is sent */ fsm_newstate(fi, CONN_STATE_SETUPWAIT); conn->path = iucv_path_alloc(NETIUCV_QUEUELEN_DEFAULT, 0, GFP_KERNEL); IUCV_DBF_TEXT_(setup, 2, "%s: connecting to %s ...\n", netdev->name, netiucv_printuser(conn)); rc = iucv_path_connect(conn->path, &netiucv_handler, conn->userid, NULL, conn->userdata, conn); switch (rc) { case 0: netdev->tx_queue_len = conn->path->msglim; fsm_addtimer(&conn->timer, NETIUCV_TIMEOUT_5SEC, CONN_EVENT_TIMER, conn); return; case 11: dev_warn(privptr->dev, "The IUCV device failed to connect to z/VM guest %s\n", netiucv_printname(conn->userid, 8)); fsm_newstate(fi, CONN_STATE_STARTWAIT); break; case 12: dev_warn(privptr->dev, "The IUCV device failed to connect to the peer on z/VM" " guest %s\n", netiucv_printname(conn->userid, 8)); fsm_newstate(fi, CONN_STATE_STARTWAIT); break; case 13: dev_err(privptr->dev, "Connecting the IUCV device would exceed the maximum" " number of IUCV connections\n"); fsm_newstate(fi, CONN_STATE_CONNERR); break; case 14: dev_err(privptr->dev, "z/VM guest %s has too many IUCV connections" " to connect with the IUCV device\n", netiucv_printname(conn->userid, 8)); fsm_newstate(fi, CONN_STATE_CONNERR); break; case 15: dev_err(privptr->dev, "The IUCV device cannot connect to a z/VM guest with no" " IUCV authorization\n"); fsm_newstate(fi, CONN_STATE_CONNERR); break; default: dev_err(privptr->dev, "Connecting the IUCV device failed with error %d\n", rc); fsm_newstate(fi, CONN_STATE_CONNERR); break; } IUCV_DBF_TEXT_(setup, 5, "iucv_connect rc is %d\n", rc); kfree(conn->path); conn->path = NULL; } static void netiucv_purge_skb_queue(struct sk_buff_head *q) { struct sk_buff *skb; while ((skb = skb_dequeue(q))) { atomic_dec(&skb->users); dev_kfree_skb_any(skb); } } static void conn_action_stop(fsm_instance *fi, int event, void *arg) { struct iucv_event *ev = arg; struct iucv_connection *conn = ev->conn; struct net_device *netdev = conn->netdev; struct netiucv_priv *privptr = netdev_priv(netdev); IUCV_DBF_TEXT(trace, 3, __func__); fsm_deltimer(&conn->timer); fsm_newstate(fi, CONN_STATE_STOPPED); netiucv_purge_skb_queue(&conn->collect_queue); if (conn->path) { IUCV_DBF_TEXT(trace, 5, "calling iucv_path_sever\n"); iucv_path_sever(conn->path, conn->userdata); kfree(conn->path); conn->path = NULL; } netiucv_purge_skb_queue(&conn->commit_queue); fsm_event(privptr->fsm, DEV_EVENT_CONDOWN, netdev); } static void conn_action_inval(fsm_instance *fi, int event, void *arg) { struct iucv_connection *conn = arg; struct net_device *netdev = conn->netdev; IUCV_DBF_TEXT_(data, 2, "%s('%s'): conn_action_inval called\n", netdev->name, conn->userid); } static const fsm_node conn_fsm[] = { { CONN_STATE_INVALID, CONN_EVENT_START, conn_action_inval }, { CONN_STATE_STOPPED, CONN_EVENT_START, conn_action_start }, { CONN_STATE_STOPPED, CONN_EVENT_STOP, conn_action_stop }, { CONN_STATE_STARTWAIT, CONN_EVENT_STOP, conn_action_stop }, { CONN_STATE_SETUPWAIT, CONN_EVENT_STOP, conn_action_stop }, { CONN_STATE_IDLE, CONN_EVENT_STOP, conn_action_stop }, { CONN_STATE_TX, CONN_EVENT_STOP, conn_action_stop }, { CONN_STATE_REGERR, CONN_EVENT_STOP, conn_action_stop }, { CONN_STATE_CONNERR, CONN_EVENT_STOP, conn_action_stop }, { CONN_STATE_STOPPED, CONN_EVENT_CONN_REQ, conn_action_connreject }, { CONN_STATE_STARTWAIT, CONN_EVENT_CONN_REQ, conn_action_connaccept }, { CONN_STATE_SETUPWAIT, CONN_EVENT_CONN_REQ, conn_action_connaccept }, { CONN_STATE_IDLE, CONN_EVENT_CONN_REQ, conn_action_connreject }, { CONN_STATE_TX, CONN_EVENT_CONN_REQ, conn_action_connreject }, { CONN_STATE_SETUPWAIT, CONN_EVENT_CONN_ACK, conn_action_connack }, { CONN_STATE_SETUPWAIT, CONN_EVENT_TIMER, conn_action_conntimsev }, { CONN_STATE_SETUPWAIT, CONN_EVENT_CONN_REJ, conn_action_connsever }, { CONN_STATE_IDLE, CONN_EVENT_CONN_REJ, conn_action_connsever }, { CONN_STATE_TX, CONN_EVENT_CONN_REJ, conn_action_connsever }, { CONN_STATE_IDLE, CONN_EVENT_RX, conn_action_rx }, { CONN_STATE_TX, CONN_EVENT_RX, conn_action_rx }, { CONN_STATE_TX, CONN_EVENT_TXDONE, conn_action_txdone }, { CONN_STATE_IDLE, CONN_EVENT_TXDONE, conn_action_txdone }, }; static const int CONN_FSM_LEN = sizeof(conn_fsm) / sizeof(fsm_node); /* * Actions for interface - statemachine. */ /** * dev_action_start * @fi: An instance of an interface statemachine. * @event: The event, just happened. * @arg: Generic pointer, casted from struct net_device * upon call. * * Startup connection by sending CONN_EVENT_START to it. */ static void dev_action_start(fsm_instance *fi, int event, void *arg) { struct net_device *dev = arg; struct netiucv_priv *privptr = netdev_priv(dev); IUCV_DBF_TEXT(trace, 3, __func__); fsm_newstate(fi, DEV_STATE_STARTWAIT); fsm_event(privptr->conn->fsm, CONN_EVENT_START, privptr->conn); } /** * Shutdown connection by sending CONN_EVENT_STOP to it. * * @param fi An instance of an interface statemachine. * @param event The event, just happened. * @param arg Generic pointer, casted from struct net_device * upon call. */ static void dev_action_stop(fsm_instance *fi, int event, void *arg) { struct net_device *dev = arg; struct netiucv_priv *privptr = netdev_priv(dev); struct iucv_event ev; IUCV_DBF_TEXT(trace, 3, __func__); ev.conn = privptr->conn; fsm_newstate(fi, DEV_STATE_STOPWAIT); fsm_event(privptr->conn->fsm, CONN_EVENT_STOP, &ev); } /** * Called from connection statemachine * when a connection is up and running. * * @param fi An instance of an interface statemachine. * @param event The event, just happened. * @param arg Generic pointer, casted from struct net_device * upon call. */ static void dev_action_connup(fsm_instance *fi, int event, void *arg) { struct net_device *dev = arg; struct netiucv_priv *privptr = netdev_priv(dev); IUCV_DBF_TEXT(trace, 3, __func__); switch (fsm_getstate(fi)) { case DEV_STATE_STARTWAIT: fsm_newstate(fi, DEV_STATE_RUNNING); dev_info(privptr->dev, "The IUCV device has been connected" " successfully to %s\n", netiucv_printuser(privptr->conn)); IUCV_DBF_TEXT(setup, 3, "connection is up and running\n"); break; case DEV_STATE_STOPWAIT: IUCV_DBF_TEXT(data, 2, "dev_action_connup: in DEV_STATE_STOPWAIT\n"); break; } } /** * Called from connection statemachine * when a connection has been shutdown. * * @param fi An instance of an interface statemachine. * @param event The event, just happened. * @param arg Generic pointer, casted from struct net_device * upon call. */ static void dev_action_conndown(fsm_instance *fi, int event, void *arg) { IUCV_DBF_TEXT(trace, 3, __func__); switch (fsm_getstate(fi)) { case DEV_STATE_RUNNING: fsm_newstate(fi, DEV_STATE_STARTWAIT); break; case DEV_STATE_STOPWAIT: fsm_newstate(fi, DEV_STATE_STOPPED); IUCV_DBF_TEXT(setup, 3, "connection is down\n"); break; } } static const fsm_node dev_fsm[] = { { DEV_STATE_STOPPED, DEV_EVENT_START, dev_action_start }, { DEV_STATE_STOPWAIT, DEV_EVENT_START, dev_action_start }, { DEV_STATE_STOPWAIT, DEV_EVENT_CONDOWN, dev_action_conndown }, { DEV_STATE_STARTWAIT, DEV_EVENT_STOP, dev_action_stop }, { DEV_STATE_STARTWAIT, DEV_EVENT_CONUP, dev_action_connup }, { DEV_STATE_RUNNING, DEV_EVENT_STOP, dev_action_stop }, { DEV_STATE_RUNNING, DEV_EVENT_CONDOWN, dev_action_conndown }, { DEV_STATE_RUNNING, DEV_EVENT_CONUP, netiucv_action_nop }, }; static const int DEV_FSM_LEN = sizeof(dev_fsm) / sizeof(fsm_node); /** * Transmit a packet. * This is a helper function for netiucv_tx(). * * @param conn Connection to be used for sending. * @param skb Pointer to struct sk_buff of packet to send. * The linklevel header has already been set up * by netiucv_tx(). * * @return 0 on success, -ERRNO on failure. (Never fails.) */ static int netiucv_transmit_skb(struct iucv_connection *conn, struct sk_buff *skb) { struct iucv_message msg; unsigned long saveflags; struct ll_header header; int rc; if (fsm_getstate(conn->fsm) != CONN_STATE_IDLE) { int l = skb->len + NETIUCV_HDRLEN; spin_lock_irqsave(&conn->collect_lock, saveflags); if (conn->collect_len + l > (conn->max_buffsize - NETIUCV_HDRLEN)) { rc = -EBUSY; IUCV_DBF_TEXT(data, 2, "EBUSY from netiucv_transmit_skb\n"); } else { atomic_inc(&skb->users); skb_queue_tail(&conn->collect_queue, skb); conn->collect_len += l; rc = 0; } spin_unlock_irqrestore(&conn->collect_lock, saveflags); } else { struct sk_buff *nskb = skb; /** * Copy the skb to a new allocated skb in lowmem only if the * data is located above 2G in memory or tailroom is < 2. */ unsigned long hi = ((unsigned long)(skb_tail_pointer(skb) + NETIUCV_HDRLEN)) >> 31; int copied = 0; if (hi || (skb_tailroom(skb) < 2)) { nskb = alloc_skb(skb->len + NETIUCV_HDRLEN + NETIUCV_HDRLEN, GFP_ATOMIC | GFP_DMA); if (!nskb) { IUCV_DBF_TEXT(data, 2, "alloc_skb failed\n"); rc = -ENOMEM; return rc; } else { skb_reserve(nskb, NETIUCV_HDRLEN); memcpy(skb_put(nskb, skb->len), skb->data, skb->len); } copied = 1; } /** * skb now is below 2G and has enough room. Add headers. */ header.next = nskb->len + NETIUCV_HDRLEN; memcpy(skb_push(nskb, NETIUCV_HDRLEN), &header, NETIUCV_HDRLEN); header.next = 0; memcpy(skb_put(nskb, NETIUCV_HDRLEN), &header, NETIUCV_HDRLEN); fsm_newstate(conn->fsm, CONN_STATE_TX); conn->prof.send_stamp = current_kernel_time(); msg.tag = 1; msg.class = 0; rc = iucv_message_send(conn->path, &msg, 0, 0, nskb->data, nskb->len); conn->prof.doios_single++; conn->prof.txlen += skb->len; conn->prof.tx_pending++; if (conn->prof.tx_pending > conn->prof.tx_max_pending) conn->prof.tx_max_pending = conn->prof.tx_pending; if (rc) { struct netiucv_priv *privptr; fsm_newstate(conn->fsm, CONN_STATE_IDLE); conn->prof.tx_pending--; privptr = netdev_priv(conn->netdev); if (privptr) privptr->stats.tx_errors++; if (copied) dev_kfree_skb(nskb); else { /** * Remove our headers. They get added * again on retransmit. */ skb_pull(skb, NETIUCV_HDRLEN); skb_trim(skb, skb->len - NETIUCV_HDRLEN); } IUCV_DBF_TEXT_(data, 2, "rc %d from iucv_send\n", rc); } else { if (copied) dev_kfree_skb(skb); atomic_inc(&nskb->users); skb_queue_tail(&conn->commit_queue, nskb); } } return rc; } /* * Interface API for upper network layers */ /** * Open an interface. * Called from generic network layer when ifconfig up is run. * * @param dev Pointer to interface struct. * * @return 0 on success, -ERRNO on failure. (Never fails.) */ static int netiucv_open(struct net_device *dev) { struct netiucv_priv *priv = netdev_priv(dev); fsm_event(priv->fsm, DEV_EVENT_START, dev); return 0; } /** * Close an interface. * Called from generic network layer when ifconfig down is run. * * @param dev Pointer to interface struct. * * @return 0 on success, -ERRNO on failure. (Never fails.) */ static int netiucv_close(struct net_device *dev) { struct netiucv_priv *priv = netdev_priv(dev); fsm_event(priv->fsm, DEV_EVENT_STOP, dev); return 0; } static int netiucv_pm_prepare(struct device *dev) { IUCV_DBF_TEXT(trace, 3, __func__); return 0; } static void netiucv_pm_complete(struct device *dev) { IUCV_DBF_TEXT(trace, 3, __func__); return; } /** * netiucv_pm_freeze() - Freeze PM callback * @dev: netiucv device * * close open netiucv interfaces */ static int netiucv_pm_freeze(struct device *dev) { struct netiucv_priv *priv = dev_get_drvdata(dev); struct net_device *ndev = NULL; int rc = 0; IUCV_DBF_TEXT(trace, 3, __func__); if (priv && priv->conn) ndev = priv->conn->netdev; if (!ndev) goto out; netif_device_detach(ndev); priv->pm_state = fsm_getstate(priv->fsm); rc = netiucv_close(ndev); out: return rc; } /** * netiucv_pm_restore_thaw() - Thaw and restore PM callback * @dev: netiucv device * * re-open netiucv interfaces closed during freeze */ static int netiucv_pm_restore_thaw(struct device *dev) { struct netiucv_priv *priv = dev_get_drvdata(dev); struct net_device *ndev = NULL; int rc = 0; IUCV_DBF_TEXT(trace, 3, __func__); if (priv && priv->conn) ndev = priv->conn->netdev; if (!ndev) goto out; switch (priv->pm_state) { case DEV_STATE_RUNNING: case DEV_STATE_STARTWAIT: rc = netiucv_open(ndev); break; default: break; } netif_device_attach(ndev); out: return rc; } /** * Start transmission of a packet. * Called from generic network device layer. * * @param skb Pointer to buffer containing the packet. * @param dev Pointer to interface struct. * * @return 0 if packet consumed, !0 if packet rejected. * Note: If we return !0, then the packet is free'd by * the generic network layer. */ static int netiucv_tx(struct sk_buff *skb, struct net_device *dev) { struct netiucv_priv *privptr = netdev_priv(dev); int rc; IUCV_DBF_TEXT(trace, 4, __func__); /** * Some sanity checks ... */ if (skb == NULL) { IUCV_DBF_TEXT(data, 2, "netiucv_tx: skb is NULL\n"); privptr->stats.tx_dropped++; return NETDEV_TX_OK; } if (skb_headroom(skb) < NETIUCV_HDRLEN) { IUCV_DBF_TEXT(data, 2, "netiucv_tx: skb_headroom < NETIUCV_HDRLEN\n"); dev_kfree_skb(skb); privptr->stats.tx_dropped++; return NETDEV_TX_OK; } /** * If connection is not running, try to restart it * and throw away packet. */ if (fsm_getstate(privptr->fsm) != DEV_STATE_RUNNING) { dev_kfree_skb(skb); privptr->stats.tx_dropped++; privptr->stats.tx_errors++; privptr->stats.tx_carrier_errors++; return NETDEV_TX_OK; } if (netiucv_test_and_set_busy(dev)) { IUCV_DBF_TEXT(data, 2, "EBUSY from netiucv_tx\n"); return NETDEV_TX_BUSY; } dev->trans_start = jiffies; rc = netiucv_transmit_skb(privptr->conn, skb); netiucv_clear_busy(dev); return rc ? NETDEV_TX_BUSY : NETDEV_TX_OK; } /** * netiucv_stats * @dev: Pointer to interface struct. * * Returns interface statistics of a device. * * Returns pointer to stats struct of this interface. */ static struct net_device_stats *netiucv_stats (struct net_device * dev) { struct netiucv_priv *priv = netdev_priv(dev); IUCV_DBF_TEXT(trace, 5, __func__); return &priv->stats; } /** * netiucv_change_mtu * @dev: Pointer to interface struct. * @new_mtu: The new MTU to use for this interface. * * Sets MTU of an interface. * * Returns 0 on success, -EINVAL if MTU is out of valid range. * (valid range is 576 .. NETIUCV_MTU_MAX). */ static int netiucv_change_mtu(struct net_device * dev, int new_mtu) { IUCV_DBF_TEXT(trace, 3, __func__); if (new_mtu < 576 || new_mtu > NETIUCV_MTU_MAX) { IUCV_DBF_TEXT(setup, 2, "given MTU out of valid range\n"); return -EINVAL; } dev->mtu = new_mtu; return 0; } /* * attributes in sysfs */ static ssize_t user_show(struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%s\n", netiucv_printuser(priv->conn)); } static int netiucv_check_user(const char *buf, size_t count, char *username, char *userdata) { const char *p; int i; p = strchr(buf, '.'); if ((p && ((count > 26) || ((p - buf) > 8) || (buf + count - p > 18))) || (!p && (count > 9))) { IUCV_DBF_TEXT(setup, 2, "conn_write: too long\n"); return -EINVAL; } for (i = 0, p = buf; i < 8 && *p && *p != '.'; i++, p++) { if (isalnum(*p) || *p == '$') { username[i] = toupper(*p); continue; } if (*p == '\n') /* trailing lf, grr */ break; IUCV_DBF_TEXT_(setup, 2, "conn_write: invalid character %02x\n", *p); return -EINVAL; } while (i < 8) username[i++] = ' '; username[8] = '\0'; if (*p == '.') { p++; for (i = 0; i < 16 && *p; i++, p++) { if (*p == '\n') break; userdata[i] = toupper(*p); } while (i > 0 && i < 16) userdata[i++] = ' '; } else memcpy(userdata, iucvMagic_ascii, 16); userdata[16] = '\0'; ASCEBC(userdata, 16); return 0; } static ssize_t user_write(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); struct net_device *ndev = priv->conn->netdev; char username[9]; char userdata[17]; int rc; struct iucv_connection *cp; IUCV_DBF_TEXT(trace, 3, __func__); rc = netiucv_check_user(buf, count, username, userdata); if (rc) return rc; if (memcmp(username, priv->conn->userid, 9) && (ndev->flags & (IFF_UP | IFF_RUNNING))) { /* username changed while the interface is active. */ IUCV_DBF_TEXT(setup, 2, "user_write: device active\n"); return -EPERM; } read_lock_bh(&iucv_connection_rwlock); list_for_each_entry(cp, &iucv_connection_list, list) { if (!strncmp(username, cp->userid, 9) && !strncmp(userdata, cp->userdata, 17) && cp->netdev != ndev) { read_unlock_bh(&iucv_connection_rwlock); IUCV_DBF_TEXT_(setup, 2, "user_write: Connection to %s " "already exists\n", netiucv_printuser(cp)); return -EEXIST; } } read_unlock_bh(&iucv_connection_rwlock); memcpy(priv->conn->userid, username, 9); memcpy(priv->conn->userdata, userdata, 17); return count; } static DEVICE_ATTR(user, 0644, user_show, user_write); static ssize_t buffer_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%d\n", priv->conn->max_buffsize); } static ssize_t buffer_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); struct net_device *ndev = priv->conn->netdev; char *e; int bs1; IUCV_DBF_TEXT(trace, 3, __func__); if (count >= 39) return -EINVAL; bs1 = simple_strtoul(buf, &e, 0); if (e && (!isspace(*e))) { IUCV_DBF_TEXT_(setup, 2, "buffer_write: invalid char %02x\n", *e); return -EINVAL; } if (bs1 > NETIUCV_BUFSIZE_MAX) { IUCV_DBF_TEXT_(setup, 2, "buffer_write: buffer size %d too large\n", bs1); return -EINVAL; } if ((ndev->flags & IFF_RUNNING) && (bs1 < (ndev->mtu + NETIUCV_HDRLEN + 2))) { IUCV_DBF_TEXT_(setup, 2, "buffer_write: buffer size %d too small\n", bs1); return -EINVAL; } if (bs1 < (576 + NETIUCV_HDRLEN + NETIUCV_HDRLEN)) { IUCV_DBF_TEXT_(setup, 2, "buffer_write: buffer size %d too small\n", bs1); return -EINVAL; } priv->conn->max_buffsize = bs1; if (!(ndev->flags & IFF_RUNNING)) ndev->mtu = bs1 - NETIUCV_HDRLEN - NETIUCV_HDRLEN; return count; } static DEVICE_ATTR(buffer, 0644, buffer_show, buffer_write); static ssize_t dev_fsm_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%s\n", fsm_getstate_str(priv->fsm)); } static DEVICE_ATTR(device_fsm_state, 0444, dev_fsm_show, NULL); static ssize_t conn_fsm_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%s\n", fsm_getstate_str(priv->conn->fsm)); } static DEVICE_ATTR(connection_fsm_state, 0444, conn_fsm_show, NULL); static ssize_t maxmulti_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.maxmulti); } static ssize_t maxmulti_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 4, __func__); priv->conn->prof.maxmulti = 0; return count; } static DEVICE_ATTR(max_tx_buffer_used, 0644, maxmulti_show, maxmulti_write); static ssize_t maxcq_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.maxcqueue); } static ssize_t maxcq_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 4, __func__); priv->conn->prof.maxcqueue = 0; return count; } static DEVICE_ATTR(max_chained_skbs, 0644, maxcq_show, maxcq_write); static ssize_t sdoio_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.doios_single); } static ssize_t sdoio_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 4, __func__); priv->conn->prof.doios_single = 0; return count; } static DEVICE_ATTR(tx_single_write_ops, 0644, sdoio_show, sdoio_write); static ssize_t mdoio_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.doios_multi); } static ssize_t mdoio_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); priv->conn->prof.doios_multi = 0; return count; } static DEVICE_ATTR(tx_multi_write_ops, 0644, mdoio_show, mdoio_write); static ssize_t txlen_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.txlen); } static ssize_t txlen_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 4, __func__); priv->conn->prof.txlen = 0; return count; } static DEVICE_ATTR(netto_bytes, 0644, txlen_show, txlen_write); static ssize_t txtime_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.tx_time); } static ssize_t txtime_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 4, __func__); priv->conn->prof.tx_time = 0; return count; } static DEVICE_ATTR(max_tx_io_time, 0644, txtime_show, txtime_write); static ssize_t txpend_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.tx_pending); } static ssize_t txpend_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 4, __func__); priv->conn->prof.tx_pending = 0; return count; } static DEVICE_ATTR(tx_pending, 0644, txpend_show, txpend_write); static ssize_t txmpnd_show (struct device *dev, struct device_attribute *attr, char *buf) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 5, __func__); return sprintf(buf, "%ld\n", priv->conn->prof.tx_max_pending); } static ssize_t txmpnd_write (struct device *dev, struct device_attribute *attr, const char *buf, size_t count) { struct netiucv_priv *priv = dev_get_drvdata(dev); IUCV_DBF_TEXT(trace, 4, __func__); priv->conn->prof.tx_max_pending = 0; return count; } static DEVICE_ATTR(tx_max_pending, 0644, txmpnd_show, txmpnd_write); static struct attribute *netiucv_attrs[] = { &dev_attr_buffer.attr, &dev_attr_user.attr, NULL, }; static struct attribute_group netiucv_attr_group = { .attrs = netiucv_attrs, }; static struct attribute *netiucv_stat_attrs[] = { &dev_attr_device_fsm_state.attr, &dev_attr_connection_fsm_state.attr, &dev_attr_max_tx_buffer_used.attr, &dev_attr_max_chained_skbs.attr, &dev_attr_tx_single_write_ops.attr, &dev_attr_tx_multi_write_ops.attr, &dev_attr_netto_bytes.attr, &dev_attr_max_tx_io_time.attr, &dev_attr_tx_pending.attr, &dev_attr_tx_max_pending.attr, NULL, }; static struct attribute_group netiucv_stat_attr_group = { .name = "stats", .attrs = netiucv_stat_attrs, }; static const struct attribute_group *netiucv_attr_groups[] = { &netiucv_stat_attr_group, &netiucv_attr_group, NULL, }; static int netiucv_register_device(struct net_device *ndev) { struct netiucv_priv *priv = netdev_priv(ndev); struct device *dev = kzalloc(sizeof(struct device), GFP_KERNEL); int ret; IUCV_DBF_TEXT(trace, 3, __func__); if (dev) { dev_set_name(dev, "net%s", ndev->name); dev->bus = &iucv_bus; dev->parent = iucv_root; dev->groups = netiucv_attr_groups; /* * The release function could be called after the * module has been unloaded. It's _only_ task is to * free the struct. Therefore, we specify kfree() * directly here. (Probably a little bit obfuscating * but legitime ...). */ dev->release = (void (*)(struct device *))kfree; dev->driver = &netiucv_driver; } else return -ENOMEM; ret = device_register(dev); if (ret) { put_device(dev); return ret; } priv->dev = dev; dev_set_drvdata(dev, priv); return 0; } static void netiucv_unregister_device(struct device *dev) { IUCV_DBF_TEXT(trace, 3, __func__); device_unregister(dev); } /** * Allocate and initialize a new connection structure. * Add it to the list of netiucv connections; */ static struct iucv_connection *netiucv_new_connection(struct net_device *dev, char *username, char *userdata) { struct iucv_connection *conn; conn = kzalloc(sizeof(*conn), GFP_KERNEL); if (!conn) goto out; skb_queue_head_init(&conn->collect_queue); skb_queue_head_init(&conn->commit_queue); spin_lock_init(&conn->collect_lock); conn->max_buffsize = NETIUCV_BUFSIZE_DEFAULT; conn->netdev = dev; conn->rx_buff = alloc_skb(conn->max_buffsize, GFP_KERNEL | GFP_DMA); if (!conn->rx_buff) goto out_conn; conn->tx_buff = alloc_skb(conn->max_buffsize, GFP_KERNEL | GFP_DMA); if (!conn->tx_buff) goto out_rx; conn->fsm = init_fsm("netiucvconn", conn_state_names, conn_event_names, NR_CONN_STATES, NR_CONN_EVENTS, conn_fsm, CONN_FSM_LEN, GFP_KERNEL); if (!conn->fsm) goto out_tx; fsm_settimer(conn->fsm, &conn->timer); fsm_newstate(conn->fsm, CONN_STATE_INVALID); if (userdata) memcpy(conn->userdata, userdata, 17); if (username) { memcpy(conn->userid, username, 9); fsm_newstate(conn->fsm, CONN_STATE_STOPPED); } write_lock_bh(&iucv_connection_rwlock); list_add_tail(&conn->list, &iucv_connection_list); write_unlock_bh(&iucv_connection_rwlock); return conn; out_tx: kfree_skb(conn->tx_buff); out_rx: kfree_skb(conn->rx_buff); out_conn: kfree(conn); out: return NULL; } /** * Release a connection structure and remove it from the * list of netiucv connections. */ static void netiucv_remove_connection(struct iucv_connection *conn) { IUCV_DBF_TEXT(trace, 3, __func__); write_lock_bh(&iucv_connection_rwlock); list_del_init(&conn->list); write_unlock_bh(&iucv_connection_rwlock); fsm_deltimer(&conn->timer); netiucv_purge_skb_queue(&conn->collect_queue); if (conn->path) { iucv_path_sever(conn->path, conn->userdata); kfree(conn->path); conn->path = NULL; } netiucv_purge_skb_queue(&conn->commit_queue); kfree_fsm(conn->fsm); kfree_skb(conn->rx_buff); kfree_skb(conn->tx_buff); } /** * Release everything of a net device. */ static void netiucv_free_netdevice(struct net_device *dev) { struct netiucv_priv *privptr = netdev_priv(dev); IUCV_DBF_TEXT(trace, 3, __func__); if (!dev) return; if (privptr) { if (privptr->conn) netiucv_remove_connection(privptr->conn); if (privptr->fsm) kfree_fsm(privptr->fsm); privptr->conn = NULL; privptr->fsm = NULL; /* privptr gets freed by free_netdev() */ } free_netdev(dev); } /** * Initialize a net device. (Called from kernel in alloc_netdev()) */ static const struct net_device_ops netiucv_netdev_ops = { .ndo_open = netiucv_open, .ndo_stop = netiucv_close, .ndo_get_stats = netiucv_stats, .ndo_start_xmit = netiucv_tx, .ndo_change_mtu = netiucv_change_mtu, }; static void netiucv_setup_netdevice(struct net_device *dev) { dev->mtu = NETIUCV_MTU_DEFAULT; dev->destructor = netiucv_free_netdevice; dev->hard_header_len = NETIUCV_HDRLEN; dev->addr_len = 0; dev->type = ARPHRD_SLIP; dev->tx_queue_len = NETIUCV_QUEUELEN_DEFAULT; dev->flags = IFF_POINTOPOINT | IFF_NOARP; dev->netdev_ops = &netiucv_netdev_ops; } /** * Allocate and initialize everything of a net device. */ static struct net_device *netiucv_init_netdevice(char *username, char *userdata) { struct netiucv_priv *privptr; struct net_device *dev; dev = alloc_netdev(sizeof(struct netiucv_priv), "iucv%d", netiucv_setup_netdevice); if (!dev) return NULL; rtnl_lock(); if (dev_alloc_name(dev, dev->name) < 0) goto out_netdev; privptr = netdev_priv(dev); privptr->fsm = init_fsm("netiucvdev", dev_state_names, dev_event_names, NR_DEV_STATES, NR_DEV_EVENTS, dev_fsm, DEV_FSM_LEN, GFP_KERNEL); if (!privptr->fsm) goto out_netdev; privptr->conn = netiucv_new_connection(dev, username, userdata); if (!privptr->conn) { IUCV_DBF_TEXT(setup, 2, "NULL from netiucv_new_connection\n"); goto out_fsm; } fsm_newstate(privptr->fsm, DEV_STATE_STOPPED); return dev; out_fsm: kfree_fsm(privptr->fsm); out_netdev: rtnl_unlock(); free_netdev(dev); return NULL; } static ssize_t conn_write(struct device_driver *drv, const char *buf, size_t count) { char username[9]; char userdata[17]; int rc; struct net_device *dev; struct netiucv_priv *priv; struct iucv_connection *cp; IUCV_DBF_TEXT(trace, 3, __func__); rc = netiucv_check_user(buf, count, username, userdata); if (rc) return rc; read_lock_bh(&iucv_connection_rwlock); list_for_each_entry(cp, &iucv_connection_list, list) { if (!strncmp(username, cp->userid, 9) && !strncmp(userdata, cp->userdata, 17)) { read_unlock_bh(&iucv_connection_rwlock); IUCV_DBF_TEXT_(setup, 2, "conn_write: Connection to %s " "already exists\n", netiucv_printuser(cp)); return -EEXIST; } } read_unlock_bh(&iucv_connection_rwlock); dev = netiucv_init_netdevice(username, userdata); if (!dev) { IUCV_DBF_TEXT(setup, 2, "NULL from netiucv_init_netdevice\n"); return -ENODEV; } rc = netiucv_register_device(dev); if (rc) { rtnl_unlock(); IUCV_DBF_TEXT_(setup, 2, "ret %d from netiucv_register_device\n", rc); goto out_free_ndev; } /* sysfs magic */ priv = netdev_priv(dev); SET_NETDEV_DEV(dev, priv->dev); rc = register_netdevice(dev); rtnl_unlock(); if (rc) goto out_unreg; dev_info(priv->dev, "The IUCV interface to %s has been established " "successfully\n", netiucv_printuser(priv->conn)); return count; out_unreg: netiucv_unregister_device(priv->dev); out_free_ndev: netiucv_free_netdevice(dev); return rc; } static DRIVER_ATTR(connection, 0200, NULL, conn_write); static ssize_t remove_write (struct device_driver *drv, const char *buf, size_t count) { struct iucv_connection *cp; struct net_device *ndev; struct netiucv_priv *priv; struct device *dev; char name[IFNAMSIZ]; const char *p; int i; IUCV_DBF_TEXT(trace, 3, __func__); if (count >= IFNAMSIZ) count = IFNAMSIZ - 1; for (i = 0, p = buf; i < count && *p; i++, p++) { if (*p == '\n' || *p == ' ') /* trailing lf, grr */ break; name[i] = *p; } name[i] = '\0'; read_lock_bh(&iucv_connection_rwlock); list_for_each_entry(cp, &iucv_connection_list, list) { ndev = cp->netdev; priv = netdev_priv(ndev); dev = priv->dev; if (strncmp(name, ndev->name, count)) continue; read_unlock_bh(&iucv_connection_rwlock); if (ndev->flags & (IFF_UP | IFF_RUNNING)) { dev_warn(dev, "The IUCV device is connected" " to %s and cannot be removed\n", priv->conn->userid); IUCV_DBF_TEXT(data, 2, "remove_write: still active\n"); return -EPERM; } unregister_netdev(ndev); netiucv_unregister_device(dev); return count; } read_unlock_bh(&iucv_connection_rwlock); IUCV_DBF_TEXT(data, 2, "remove_write: unknown device\n"); return -EINVAL; } static DRIVER_ATTR(remove, 0200, NULL, remove_write); static struct attribute * netiucv_drv_attrs[] = { &driver_attr_connection.attr, &driver_attr_remove.attr, NULL, }; static struct attribute_group netiucv_drv_attr_group = { .attrs = netiucv_drv_attrs, }; static const struct attribute_group *netiucv_drv_attr_groups[] = { &netiucv_drv_attr_group, NULL, }; static void netiucv_banner(void) { pr_info("driver initialized\n"); } static void __exit netiucv_exit(void) { struct iucv_connection *cp; struct net_device *ndev; struct netiucv_priv *priv; struct device *dev; IUCV_DBF_TEXT(trace, 3, __func__); while (!list_empty(&iucv_connection_list)) { cp = list_entry(iucv_connection_list.next, struct iucv_connection, list); ndev = cp->netdev; priv = netdev_priv(ndev); dev = priv->dev; unregister_netdev(ndev); netiucv_unregister_device(dev); } device_unregister(netiucv_dev); driver_unregister(&netiucv_driver); iucv_unregister(&netiucv_handler, 1); iucv_unregister_dbf_views(); pr_info("driver unloaded\n"); return; } static int __init netiucv_init(void) { int rc; rc = iucv_register_dbf_views(); if (rc) goto out; rc = iucv_register(&netiucv_handler, 1); if (rc) goto out_dbf; IUCV_DBF_TEXT(trace, 3, __func__); netiucv_driver.groups = netiucv_drv_attr_groups; rc = driver_register(&netiucv_driver); if (rc) { IUCV_DBF_TEXT_(setup, 2, "ret %d from driver_register\n", rc); goto out_iucv; } /* establish dummy device */ netiucv_dev = kzalloc(sizeof(struct device), GFP_KERNEL); if (!netiucv_dev) { rc = -ENOMEM; goto out_driver; } dev_set_name(netiucv_dev, "netiucv"); netiucv_dev->bus = &iucv_bus; netiucv_dev->parent = iucv_root; netiucv_dev->release = (void (*)(struct device *))kfree; netiucv_dev->driver = &netiucv_driver; rc = device_register(netiucv_dev); if (rc) { put_device(netiucv_dev); goto out_driver; } netiucv_banner(); return rc; out_driver: driver_unregister(&netiucv_driver); out_iucv: iucv_unregister(&netiucv_handler, 1); out_dbf: iucv_unregister_dbf_views(); out: return rc; } module_init(netiucv_init); module_exit(netiucv_exit); MODULE_LICENSE("GPL");