/* * block queue tracing application * * Copyright (C) 2005 Jens Axboe <axboe@suse.de> * Copyright (C) 2006 Jens Axboe <axboe@kernel.dk> * * Rewrite to have a single thread per CPU (managing all devices on that CPU) * Alan D. Brunelle <alan.brunelle@hp.com> - January 2009 * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA * */ #include <errno.h> #include <stdarg.h> #include <stdio.h> #include <stdlib.h> #include <string.h> #include <fcntl.h> #include <getopt.h> #include <sched.h> #include <unistd.h> #include <poll.h> #include <signal.h> #include <pthread.h> #include <locale.h> #include <sys/ioctl.h> #include <sys/types.h> #include <sys/stat.h> #include <sys/vfs.h> #include <sys/mman.h> #include <sys/param.h> #include <sys/time.h> #include <sys/resource.h> #include <sys/socket.h> #include <netinet/in.h> #include <arpa/inet.h> #include <netdb.h> #include <sys/sendfile.h> #include "btt/list.h" #include "blktrace.h" /* * You may want to increase this even more, if you are logging at a high * rate and see skipped/missed events */ #define BUF_SIZE (512 * 1024) #define BUF_NR (4) #define FILE_VBUF_SIZE (128 * 1024) #define DEBUGFS_TYPE (0x64626720) #define TRACE_NET_PORT (8462) enum { Net_none = 0, Net_server, Net_client, }; enum thread_status { Th_running, Th_leaving, Th_error }; /* * Generic stats collected: nevents can be _roughly_ estimated by data_read * (discounting pdu...) * * These fields are updated w/ pdc_dr_update & pdc_nev_update below. */ struct pdc_stats { unsigned long long data_read; unsigned long long nevents; }; struct devpath { struct list_head head; char *path; /* path to device special file */ char *buts_name; /* name returned from bt kernel code */ struct pdc_stats *stats; int fd, ncpus; unsigned long long drops; /* * For piped output only: * * Each tracer will have a tracer_devpath_head that it will add new * data onto. It's list is protected above (tracer_devpath_head.mutex) * and it will signal the processing thread using the dp_cond, * dp_mutex & dp_entries variables above. */ struct tracer_devpath_head *heads; /* * For network server mode only: */ struct cl_host *ch; u32 cl_id; time_t cl_connect_time; struct io_info *ios; }; /* * For piped output to stdout we will have each tracer thread (one per dev) * tack buffers read from the relay queues on a per-device list. * * The main thread will then collect trace buffers from each of lists in turn. * * We will use a mutex to guard each of the trace_buf list. The tracers * can then signal the main thread using <dp_cond,dp_mutex> and * dp_entries. (When dp_entries is 0, and a tracer adds an entry it will * signal. When dp_entries is 0, the main thread will wait for that condition * to be signalled.) * * adb: It may be better just to have a large buffer per tracer per dev, * and then use it as a ring-buffer. This would certainly cut down a lot * of malloc/free thrashing, at the cost of more memory movements (potentially). */ struct trace_buf { struct list_head head; struct devpath *dpp; void *buf; int cpu, len; }; struct tracer_devpath_head { pthread_mutex_t mutex; struct list_head head; struct trace_buf *prev; }; /* * Used to handle the mmap() interfaces for output file (containing traces) */ struct mmap_info { void *fs_buf; unsigned long long fs_size, fs_max_size, fs_off, fs_buf_len; unsigned long buf_size, buf_nr; int pagesize; }; /* * Each thread doing work on a (client) side of blktrace will have one * of these. The ios array contains input/output information, pfds holds * poll() data. The volatile's provide flags to/from the main executing * thread. */ struct tracer { struct list_head head; struct io_info *ios; struct pollfd *pfds; pthread_t thread; int cpu, nios; volatile int status, is_done; }; /* * networking stuff follows. we include a magic number so we know whether * to endianness convert or not. * * The len field is overloaded: * 0 - Indicates an "open" - allowing the server to set up for a dev/cpu * 1 - Indicates a "close" - Shut down connection orderly * * The cpu field is overloaded on close: it will contain the number of drops. */ struct blktrace_net_hdr { u32 magic; /* same as trace magic */ char buts_name[32]; /* trace name */ u32 cpu; /* for which cpu */ u32 max_cpus; u32 len; /* length of following trace data */ u32 cl_id; /* id for set of client per-cpu connections */ u32 buf_size; /* client buf_size for this trace */ u32 buf_nr; /* client buf_nr for this trace */ u32 page_size; /* client page_size for this trace */ }; /* * Each host encountered has one of these. The head is used to link this * on to the network server's ch_list. Connections associated with this * host are linked on conn_list, and any devices traced on that host * are connected on the devpaths list. */ struct cl_host { struct list_head head; struct list_head conn_list; struct list_head devpaths; struct net_server_s *ns; char *hostname; struct in_addr cl_in_addr; int connects, ndevs, cl_opens; }; /* * Each connection (client to server socket ('fd')) has one of these. A * back reference to the host ('ch'), and lists headers (for the host * list, and the network server conn_list) are also included. */ struct cl_conn { struct list_head ch_head, ns_head; struct cl_host *ch; int fd, ncpus; time_t connect_time; }; /* * The network server requires some poll structures to be maintained - * one per conection currently on conn_list. The nchs/ch_list values * are for each host connected to this server. The addr field is used * for scratch as new connections are established. */ struct net_server_s { struct list_head conn_list; struct list_head ch_list; struct pollfd *pfds; int listen_fd, connects, nchs; struct sockaddr_in addr; }; /* * This structure is (generically) used to providide information * for a read-to-write set of values. * * ifn & ifd represent input information * * ofn, ofd, ofp, obuf & mmap_info are used for output file (optionally). */ struct io_info { struct devpath *dpp; FILE *ofp; char *obuf; struct cl_conn *nc; /* Server network connection */ /* * mmap controlled output files */ struct mmap_info mmap_info; /* * Client network fields */ unsigned int ready; unsigned long long data_queued; /* * Input/output file descriptors & names */ int ifd, ofd; char ifn[MAXPATHLEN + 64]; char ofn[MAXPATHLEN + 64]; }; static char blktrace_version[] = "2.0.0"; /* * Linkage to blktrace helper routines (trace conversions) */ int data_is_native = -1; static int ndevs; static int max_cpus; static int ncpus; static cpu_set_t *online_cpus; static int pagesize; static int act_mask = ~0U; static int kill_running_trace; static int stop_watch; static int piped_output; static char *debugfs_path = "/sys/kernel/debug"; static char *output_name; static char *output_dir; static unsigned long buf_size = BUF_SIZE; static unsigned long buf_nr = BUF_NR; static FILE *pfp; static LIST_HEAD(devpaths); static LIST_HEAD(tracers); static volatile int done; /* * tracer threads add entries, the main thread takes them off and processes * them. These protect the dp_entries variable. */ static pthread_cond_t dp_cond = PTHREAD_COND_INITIALIZER; static pthread_mutex_t dp_mutex = PTHREAD_MUTEX_INITIALIZER; static volatile int dp_entries; /* * These synchronize master / thread interactions. */ static pthread_cond_t mt_cond = PTHREAD_COND_INITIALIZER; static pthread_mutex_t mt_mutex = PTHREAD_MUTEX_INITIALIZER; static volatile int nthreads_running; static volatile int nthreads_leaving; static volatile int nthreads_error; static volatile int tracers_run; /* * network cmd line params */ static struct sockaddr_in hostname_addr; static char hostname[MAXHOSTNAMELEN]; static int net_port = TRACE_NET_PORT; static int net_use_sendfile = 1; static int net_mode; static int *cl_fds; static int (*handle_pfds)(struct tracer *, int, int); static int (*handle_list)(struct tracer_devpath_head *, struct list_head *); #define S_OPTS "d:a:A:r:o:kw:vVb:n:D:lh:p:sI:" static struct option l_opts[] = { { .name = "dev", .has_arg = required_argument, .flag = NULL, .val = 'd' }, { .name = "input-devs", .has_arg = required_argument, .flag = NULL, .val = 'I' }, { .name = "act-mask", .has_arg = required_argument, .flag = NULL, .val = 'a' }, { .name = "set-mask", .has_arg = required_argument, .flag = NULL, .val = 'A' }, { .name = "relay", .has_arg = required_argument, .flag = NULL, .val = 'r' }, { .name = "output", .has_arg = required_argument, .flag = NULL, .val = 'o' }, { .name = "kill", .has_arg = no_argument, .flag = NULL, .val = 'k' }, { .name = "stopwatch", .has_arg = required_argument, .flag = NULL, .val = 'w' }, { .name = "version", .has_arg = no_argument, .flag = NULL, .val = 'v' }, { .name = "version", .has_arg = no_argument, .flag = NULL, .val = 'V' }, { .name = "buffer-size", .has_arg = required_argument, .flag = NULL, .val = 'b' }, { .name = "num-sub-buffers", .has_arg = required_argument, .flag = NULL, .val = 'n' }, { .name = "output-dir", .has_arg = required_argument, .flag = NULL, .val = 'D' }, { .name = "listen", .has_arg = no_argument, .flag = NULL, .val = 'l' }, { .name = "host", .has_arg = required_argument, .flag = NULL, .val = 'h' }, { .name = "port", .has_arg = required_argument, .flag = NULL, .val = 'p' }, { .name = "no-sendfile", .has_arg = no_argument, .flag = NULL, .val = 's' }, { .name = NULL, } }; static char usage_str[] = "\n\n" \ "-d <dev> | --dev=<dev>\n" \ "[ -r <debugfs path> | --relay=<debugfs path> ]\n" \ "[ -o <file> | --output=<file>]\n" \ "[ -D <dir> | --output-dir=<dir>\n" \ "[ -w <time> | --stopwatch=<time>]\n" \ "[ -a <action field> | --act-mask=<action field>]\n" \ "[ -A <action mask> | --set-mask=<action mask>]\n" \ "[ -b <size> | --buffer-size]\n" \ "[ -n <number> | --num-sub-buffers=<number>]\n" \ "[ -l | --listen]\n" \ "[ -h <hostname> | --host=<hostname>]\n" \ "[ -p <port number> | --port=<port number>]\n" \ "[ -s | --no-sendfile]\n" \ "[ -I <devs file> | --input-devs=<devs file>]\n" \ "[ -v <version> | --version]\n" \ "[ -V <version> | --version]\n" \ "\t-d Use specified device. May also be given last after options\n" \ "\t-r Path to mounted debugfs, defaults to /sys/kernel/debug\n" \ "\t-o File(s) to send output to\n" \ "\t-D Directory to prepend to output file names\n" \ "\t-w Stop after defined time, in seconds\n" \ "\t-a Only trace specified actions. See documentation\n" \ "\t-A Give trace mask as a single value. See documentation\n" \ "\t-b Sub buffer size in KiB (default 512)\n" \ "\t-n Number of sub buffers (default 4)\n" \ "\t-l Run in network listen mode (blktrace server)\n" \ "\t-h Run in network client mode, connecting to the given host\n" \ "\t-p Network port to use (default 8462)\n" \ "\t-s Make the network client NOT use sendfile() to transfer data\n" \ "\t-I Add devices found in <devs file>\n" \ "\t-v Print program version info\n" \ "\t-V Print program version info\n\n"; static void clear_events(struct pollfd *pfd) { pfd->events = 0; pfd->revents = 0; } static inline int net_client_use_sendfile(void) { return net_mode == Net_client && net_use_sendfile; } static inline int net_client_use_send(void) { return net_mode == Net_client && !net_use_sendfile; } static inline int use_tracer_devpaths(void) { return piped_output || net_client_use_send(); } static inline int in_addr_eq(struct in_addr a, struct in_addr b) { return a.s_addr == b.s_addr; } static inline void pdc_dr_update(struct devpath *dpp, int cpu, int data_read) { dpp->stats[cpu].data_read += data_read; } static inline void pdc_nev_update(struct devpath *dpp, int cpu, int nevents) { dpp->stats[cpu].nevents += nevents; } static void show_usage(char *prog) { fprintf(stderr, "Usage: %s %s", prog, usage_str); } /* * Create a timespec 'msec' milliseconds into the future */ static inline void make_timespec(struct timespec *tsp, long delta_msec) { struct timeval now; gettimeofday(&now, NULL); tsp->tv_sec = now.tv_sec; tsp->tv_nsec = 1000L * now.tv_usec; tsp->tv_nsec += (delta_msec * 1000000L); if (tsp->tv_nsec > 1000000000L) { long secs = tsp->tv_nsec / 1000000000L; tsp->tv_sec += secs; tsp->tv_nsec -= (secs * 1000000000L); } } /* * Add a timer to ensure wait ends */ static void t_pthread_cond_wait(pthread_cond_t *cond, pthread_mutex_t *mutex) { struct timespec ts; make_timespec(&ts, 50); pthread_cond_timedwait(cond, mutex, &ts); } static void unblock_tracers(void) { pthread_mutex_lock(&mt_mutex); tracers_run = 1; pthread_cond_broadcast(&mt_cond); pthread_mutex_unlock(&mt_mutex); } static void tracer_wait_unblock(struct tracer *tp) { pthread_mutex_lock(&mt_mutex); while (!tp->is_done && !tracers_run) pthread_cond_wait(&mt_cond, &mt_mutex); pthread_mutex_unlock(&mt_mutex); } static void tracer_signal_ready(struct tracer *tp, enum thread_status th_status, int status) { pthread_mutex_lock(&mt_mutex); tp->status = status; if (th_status == Th_running) nthreads_running++; else if (th_status == Th_error) nthreads_error++; else nthreads_leaving++; pthread_cond_signal(&mt_cond); pthread_mutex_unlock(&mt_mutex); } static void wait_tracers_ready(int ncpus_started) { pthread_mutex_lock(&mt_mutex); while ((nthreads_running + nthreads_error) < ncpus_started) t_pthread_cond_wait(&mt_cond, &mt_mutex); pthread_mutex_unlock(&mt_mutex); } static void wait_tracers_leaving(void) { pthread_mutex_lock(&mt_mutex); while (nthreads_leaving < nthreads_running) t_pthread_cond_wait(&mt_cond, &mt_mutex); pthread_mutex_unlock(&mt_mutex); } static void init_mmap_info(struct mmap_info *mip) { mip->buf_size = buf_size; mip->buf_nr = buf_nr; mip->pagesize = pagesize; } static void net_close_connection(int *fd) { shutdown(*fd, SHUT_RDWR); close(*fd); *fd = -1; } static void dpp_free(struct devpath *dpp) { if (dpp->stats) free(dpp->stats); if (dpp->ios) free(dpp->ios); if (dpp->path) free(dpp->path); if (dpp->buts_name) free(dpp->buts_name); free(dpp); } static int lock_on_cpu(int cpu) { cpu_set_t * cpu_mask; size_t size; cpu_mask = CPU_ALLOC(max_cpus); size = CPU_ALLOC_SIZE(max_cpus); CPU_ZERO_S(size, cpu_mask); CPU_SET_S(cpu, size, cpu_mask); if (sched_setaffinity(0, size, cpu_mask) < 0) { CPU_FREE(cpu_mask); return errno; } CPU_FREE(cpu_mask); return 0; } static int increase_limit(int resource, rlim_t increase) { struct rlimit rlim; int save_errno = errno; if (!getrlimit(resource, &rlim)) { rlim.rlim_cur += increase; if (rlim.rlim_cur >= rlim.rlim_max) rlim.rlim_max = rlim.rlim_cur + increase; if (!setrlimit(resource, &rlim)) return 1; } errno = save_errno; return 0; } static int handle_open_failure(void) { if (errno == ENFILE || errno == EMFILE) return increase_limit(RLIMIT_NOFILE, 16); return 0; } static int handle_mem_failure(size_t length) { if (errno == ENFILE) return handle_open_failure(); else if (errno == ENOMEM) return increase_limit(RLIMIT_MEMLOCK, 2 * length); return 0; } static FILE *my_fopen(const char *path, const char *mode) { FILE *fp; do { fp = fopen(path, mode); } while (fp == NULL && handle_open_failure()); return fp; } static int my_open(const char *path, int flags) { int fd; do { fd = open(path, flags); } while (fd < 0 && handle_open_failure()); return fd; } static int my_socket(int domain, int type, int protocol) { int fd; do { fd = socket(domain, type, protocol); } while (fd < 0 && handle_open_failure()); return fd; } static int my_accept(int sockfd, struct sockaddr *addr, socklen_t *addrlen) { int fd; do { fd = accept(sockfd, addr, addrlen); } while (fd < 0 && handle_open_failure()); return fd; } static void *my_mmap(void *addr, size_t length, int prot, int flags, int fd, off_t offset) { void *new; do { new = mmap(addr, length, prot, flags, fd, offset); } while (new == MAP_FAILED && handle_mem_failure(length)); return new; } static int my_mlock(struct tracer *tp, const void *addr, size_t len) { int ret, retry = 0; do { ret = mlock(addr, len); if ((retry >= 10) && tp && tp->is_done) break; retry++; } while (ret < 0 && handle_mem_failure(len)); return ret; } static int setup_mmap(int fd, unsigned int maxlen, struct mmap_info *mip, struct tracer *tp) { if (mip->fs_off + maxlen > mip->fs_buf_len) { unsigned long nr = max(16, mip->buf_nr); if (mip->fs_buf) { munlock(mip->fs_buf, mip->fs_buf_len); munmap(mip->fs_buf, mip->fs_buf_len); mip->fs_buf = NULL; } mip->fs_off = mip->fs_size & (mip->pagesize - 1); mip->fs_buf_len = (nr * mip->buf_size) - mip->fs_off; mip->fs_max_size += mip->fs_buf_len; if (ftruncate(fd, mip->fs_max_size) < 0) { perror("setup_mmap: ftruncate"); return 1; } mip->fs_buf = my_mmap(NULL, mip->fs_buf_len, PROT_WRITE, MAP_SHARED, fd, mip->fs_size - mip->fs_off); if (mip->fs_buf == MAP_FAILED) { perror("setup_mmap: mmap"); return 1; } if (my_mlock(tp, mip->fs_buf, mip->fs_buf_len) < 0) { perror("setup_mlock: mlock"); return 1; } } return 0; } static int __stop_trace(int fd) { /* * Should be stopped, don't complain if it isn't */ ioctl(fd, BLKTRACESTOP); return ioctl(fd, BLKTRACETEARDOWN); } static int write_data(char *buf, int len) { int ret; rewrite: ret = fwrite(buf, len, 1, pfp); if (ferror(pfp) || ret != 1) { if (errno == EINTR) { clearerr(pfp); goto rewrite; } if (!piped_output || (errno != EPIPE && errno != EBADF)) { fprintf(stderr, "write(%d) failed: %d/%s\n", len, errno, strerror(errno)); } goto err; } fflush(pfp); return 0; err: clearerr(pfp); return 1; } /* * Returns the number of bytes read (successfully) */ static int __net_recv_data(int fd, void *buf, unsigned int len) { unsigned int bytes_left = len; while (bytes_left && !done) { int ret = recv(fd, buf, bytes_left, MSG_WAITALL); if (ret == 0) break; else if (ret < 0) { if (errno == EAGAIN) { usleep(50); continue; } perror("server: net_recv_data: recv failed"); break; } else { buf += ret; bytes_left -= ret; } } return len - bytes_left; } static int net_recv_data(int fd, void *buf, unsigned int len) { return __net_recv_data(fd, buf, len); } /* * Returns number of bytes written */ static int net_send_data(int fd, void *buf, unsigned int buf_len) { int ret; unsigned int bytes_left = buf_len; while (bytes_left) { ret = send(fd, buf, bytes_left, 0); if (ret < 0) { perror("send"); break; } buf += ret; bytes_left -= ret; } return buf_len - bytes_left; } static int net_send_header(int fd, int cpu, char *buts_name, int len) { struct blktrace_net_hdr hdr; memset(&hdr, 0, sizeof(hdr)); hdr.magic = BLK_IO_TRACE_MAGIC; memset(hdr.buts_name, 0, sizeof(hdr.buts_name)); strncpy(hdr.buts_name, buts_name, sizeof(hdr.buts_name)); hdr.buts_name[sizeof(hdr.buts_name) - 1] = '\0'; hdr.cpu = cpu; hdr.max_cpus = max_cpus; hdr.len = len; hdr.cl_id = getpid(); hdr.buf_size = buf_size; hdr.buf_nr = buf_nr; hdr.page_size = pagesize; return net_send_data(fd, &hdr, sizeof(hdr)) != sizeof(hdr); } static void net_send_open_close(int fd, int cpu, char *buts_name, int len) { struct blktrace_net_hdr ret_hdr; net_send_header(fd, cpu, buts_name, len); net_recv_data(fd, &ret_hdr, sizeof(ret_hdr)); } static void net_send_open(int fd, int cpu, char *buts_name) { net_send_open_close(fd, cpu, buts_name, 0); } static void net_send_close(int fd, char *buts_name, int drops) { /* * Overload CPU w/ number of drops * * XXX: Need to clear/set done around call - done=1 (which * is true here) stops reads from happening... :-( */ done = 0; net_send_open_close(fd, drops, buts_name, 1); done = 1; } static void ack_open_close(int fd, char *buts_name) { net_send_header(fd, 0, buts_name, 2); } static void net_send_drops(int fd) { struct list_head *p; __list_for_each(p, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); net_send_close(fd, dpp->buts_name, dpp->drops); } } /* * Returns: * 0: "EOF" * 1: OK * -1: Error */ static int net_get_header(struct cl_conn *nc, struct blktrace_net_hdr *bnh) { int bytes_read; int fl = fcntl(nc->fd, F_GETFL); fcntl(nc->fd, F_SETFL, fl | O_NONBLOCK); bytes_read = __net_recv_data(nc->fd, bnh, sizeof(*bnh)); fcntl(nc->fd, F_SETFL, fl & ~O_NONBLOCK); if (bytes_read == sizeof(*bnh)) return 1; else if (bytes_read == 0) return 0; else return -1; } static int net_setup_addr(void) { struct sockaddr_in *addr = &hostname_addr; memset(addr, 0, sizeof(*addr)); addr->sin_family = AF_INET; addr->sin_port = htons(net_port); if (inet_aton(hostname, &addr->sin_addr) != 1) { struct hostent *hent; retry: hent = gethostbyname(hostname); if (!hent) { if (h_errno == TRY_AGAIN) { usleep(100); goto retry; } else if (h_errno == NO_RECOVERY) { fprintf(stderr, "gethostbyname(%s)" "non-recoverable error encountered\n", hostname); } else { /* * HOST_NOT_FOUND, NO_ADDRESS or NO_DATA */ fprintf(stderr, "Host %s not found\n", hostname); } return 1; } memcpy(&addr->sin_addr, hent->h_addr, 4); memset(hostname, 0, sizeof(hostname)); strncpy(hostname, hent->h_name, sizeof(hostname)); hostname[sizeof(hostname) - 1] = '\0'; } return 0; } static int net_setup_client(void) { int fd; struct sockaddr_in *addr = &hostname_addr; fd = my_socket(AF_INET, SOCK_STREAM, 0); if (fd < 0) { perror("client: socket"); return -1; } if (connect(fd, (struct sockaddr *)addr, sizeof(*addr)) < 0) { if (errno == ECONNREFUSED) fprintf(stderr, "\nclient: Connection to %s refused, " "perhaps the server is not started?\n\n", hostname); else perror("client: connect"); close(fd); return -1; } return fd; } static int open_client_connections(void) { int cpu; size_t alloc_size = CPU_ALLOC_SIZE(max_cpus); cl_fds = calloc(ncpus, sizeof(*cl_fds)); for (cpu = 0; cpu < max_cpus; cpu++) { if (!CPU_ISSET_S(cpu, alloc_size, online_cpus)) continue; cl_fds[cpu] = net_setup_client(); if (cl_fds[cpu] < 0) goto err; } return 0; err: while (cpu > 0) close(cl_fds[cpu--]); free(cl_fds); return 1; } static void close_client_connections(void) { if (cl_fds) { int cpu, *fdp; size_t alloc_size = CPU_ALLOC_SIZE(max_cpus); for (cpu = 0, fdp = cl_fds; cpu < max_cpus; cpu++, fdp++) { if (!CPU_ISSET_S(cpu, alloc_size, online_cpus)) continue; if (*fdp >= 0) { net_send_drops(*fdp); net_close_connection(fdp); } } free(cl_fds); } } static int setup_buts(void) { struct list_head *p; int ret = 0; __list_for_each(p, &devpaths) { struct blk_user_trace_setup buts; struct devpath *dpp = list_entry(p, struct devpath, head); memset(&buts, 0, sizeof(buts)); buts.buf_size = buf_size; buts.buf_nr = buf_nr; buts.act_mask = act_mask; if (ioctl(dpp->fd, BLKTRACESETUP, &buts) >= 0) { dpp->ncpus = max_cpus; dpp->buts_name = strdup(buts.name); if (dpp->stats) free(dpp->stats); dpp->stats = calloc(dpp->ncpus, sizeof(*dpp->stats)); memset(dpp->stats, 0, dpp->ncpus * sizeof(*dpp->stats)); } else { fprintf(stderr, "BLKTRACESETUP(2) %s failed: %d/%s\n", dpp->path, errno, strerror(errno)); ret++; } } return ret; } static void start_buts(void) { struct list_head *p; __list_for_each(p, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); if (ioctl(dpp->fd, BLKTRACESTART) < 0) { fprintf(stderr, "BLKTRACESTART %s failed: %d/%s\n", dpp->path, errno, strerror(errno)); } } } static int get_drops(struct devpath *dpp) { int fd, drops = 0; char fn[MAXPATHLEN + 64], tmp[256]; snprintf(fn, sizeof(fn), "%s/block/%s/dropped", debugfs_path, dpp->buts_name); fd = my_open(fn, O_RDONLY); if (fd < 0) { /* * This may be ok: the kernel may not support * dropped counts. */ if (errno != ENOENT) fprintf(stderr, "Could not open %s: %d/%s\n", fn, errno, strerror(errno)); return 0; } else if (read(fd, tmp, sizeof(tmp)) < 0) { fprintf(stderr, "Could not read %s: %d/%s\n", fn, errno, strerror(errno)); } else drops = atoi(tmp); close(fd); return drops; } static void get_all_drops(void) { struct list_head *p; __list_for_each(p, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); dpp->drops = get_drops(dpp); } } static inline struct trace_buf *alloc_trace_buf(int cpu, int bufsize) { struct trace_buf *tbp; tbp = malloc(sizeof(*tbp) + bufsize); INIT_LIST_HEAD(&tbp->head); tbp->len = 0; tbp->buf = (void *)(tbp + 1); tbp->cpu = cpu; tbp->dpp = NULL; /* Will be set when tbp is added */ return tbp; } static void free_tracer_heads(struct devpath *dpp) { int cpu; struct tracer_devpath_head *hd; for (cpu = 0, hd = dpp->heads; cpu < max_cpus; cpu++, hd++) { if (hd->prev) free(hd->prev); pthread_mutex_destroy(&hd->mutex); } free(dpp->heads); } static int setup_tracer_devpaths(void) { struct list_head *p; if (net_client_use_send()) if (open_client_connections()) return 1; __list_for_each(p, &devpaths) { int cpu; struct tracer_devpath_head *hd; struct devpath *dpp = list_entry(p, struct devpath, head); dpp->heads = calloc(max_cpus, sizeof(struct tracer_devpath_head)); for (cpu = 0, hd = dpp->heads; cpu < max_cpus; cpu++, hd++) { INIT_LIST_HEAD(&hd->head); pthread_mutex_init(&hd->mutex, NULL); hd->prev = NULL; } } return 0; } static inline void add_trace_buf(struct devpath *dpp, int cpu, struct trace_buf **tbpp) { struct trace_buf *tbp = *tbpp; struct tracer_devpath_head *hd = &dpp->heads[cpu]; tbp->dpp = dpp; pthread_mutex_lock(&hd->mutex); list_add_tail(&tbp->head, &hd->head); pthread_mutex_unlock(&hd->mutex); *tbpp = alloc_trace_buf(cpu, buf_size); } static inline void incr_entries(int entries_handled) { pthread_mutex_lock(&dp_mutex); if (dp_entries == 0) pthread_cond_signal(&dp_cond); dp_entries += entries_handled; pthread_mutex_unlock(&dp_mutex); } static void decr_entries(int handled) { pthread_mutex_lock(&dp_mutex); dp_entries -= handled; pthread_mutex_unlock(&dp_mutex); } static int wait_empty_entries(void) { pthread_mutex_lock(&dp_mutex); while (!done && dp_entries == 0) t_pthread_cond_wait(&dp_cond, &dp_mutex); pthread_mutex_unlock(&dp_mutex); return !done; } static int add_devpath(char *path) { int fd; struct devpath *dpp; struct list_head *p; /* * Verify device is not duplicated */ __list_for_each(p, &devpaths) { struct devpath *tmp = list_entry(p, struct devpath, head); if (!strcmp(tmp->path, path)) return 0; } /* * Verify device is valid before going too far */ fd = my_open(path, O_RDONLY | O_NONBLOCK); if (fd < 0) { fprintf(stderr, "Invalid path %s specified: %d/%s\n", path, errno, strerror(errno)); return 1; } dpp = malloc(sizeof(*dpp)); memset(dpp, 0, sizeof(*dpp)); dpp->path = strdup(path); dpp->fd = fd; ndevs++; list_add_tail(&dpp->head, &devpaths); return 0; } static void rel_devpaths(void) { struct list_head *p, *q; list_for_each_safe(p, q, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); list_del(&dpp->head); __stop_trace(dpp->fd); close(dpp->fd); if (dpp->heads) free_tracer_heads(dpp); dpp_free(dpp); ndevs--; } } static int flush_subbuf_net(struct trace_buf *tbp) { int fd = cl_fds[tbp->cpu]; struct devpath *dpp = tbp->dpp; if (net_send_header(fd, tbp->cpu, dpp->buts_name, tbp->len)) return 1; else if (net_send_data(fd, tbp->buf, tbp->len) != tbp->len) return 1; return 0; } static int handle_list_net(__attribute__((__unused__))struct tracer_devpath_head *hd, struct list_head *list) { struct trace_buf *tbp; struct list_head *p, *q; int entries_handled = 0; list_for_each_safe(p, q, list) { tbp = list_entry(p, struct trace_buf, head); list_del(&tbp->head); entries_handled++; if (cl_fds[tbp->cpu] >= 0) { if (flush_subbuf_net(tbp)) { close(cl_fds[tbp->cpu]); cl_fds[tbp->cpu] = -1; } } free(tbp); } return entries_handled; } /* * Tack 'tbp's buf onto the tail of 'prev's buf */ static struct trace_buf *tb_combine(struct trace_buf *prev, struct trace_buf *tbp) { unsigned long tot_len; tot_len = prev->len + tbp->len; if (tot_len > buf_size) { /* * tbp->head isn't connected (it was 'prev' * so it had been taken off of the list * before). Therefore, we can realloc * the whole structures, as the other fields * are "static". */ prev = realloc(prev, sizeof(*prev) + tot_len); prev->buf = (void *)(prev + 1); } memcpy(prev->buf + prev->len, tbp->buf, tbp->len); prev->len = tot_len; free(tbp); return prev; } static int handle_list_file(struct tracer_devpath_head *hd, struct list_head *list) { int off, t_len, nevents; struct blk_io_trace *t; struct list_head *p, *q; int entries_handled = 0; struct trace_buf *tbp, *prev; prev = hd->prev; list_for_each_safe(p, q, list) { tbp = list_entry(p, struct trace_buf, head); list_del(&tbp->head); entries_handled++; /* * If there was some leftover before, tack this new * entry onto the tail of the previous one. */ if (prev) tbp = tb_combine(prev, tbp); /* * See how many whole traces there are - send them * all out in one go. */ off = 0; nevents = 0; while (off + (int)sizeof(*t) <= tbp->len) { t = (struct blk_io_trace *)(tbp->buf + off); t_len = sizeof(*t) + t->pdu_len; if (off + t_len > tbp->len) break; off += t_len; nevents++; } if (nevents) pdc_nev_update(tbp->dpp, tbp->cpu, nevents); /* * Write any full set of traces, any remaining data is kept * for the next pass. */ if (off) { if (write_data(tbp->buf, off) || off == tbp->len) { free(tbp); prev = NULL; } else { /* * Move valid data to beginning of buffer */ tbp->len -= off; memmove(tbp->buf, tbp->buf + off, tbp->len); prev = tbp; } } else prev = tbp; } hd->prev = prev; return entries_handled; } static void __process_trace_bufs(void) { int cpu; struct list_head *p; struct list_head list; int handled = 0; __list_for_each(p, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); struct tracer_devpath_head *hd = dpp->heads; for (cpu = 0; cpu < max_cpus; cpu++, hd++) { pthread_mutex_lock(&hd->mutex); if (list_empty(&hd->head)) { pthread_mutex_unlock(&hd->mutex); continue; } list_replace_init(&hd->head, &list); pthread_mutex_unlock(&hd->mutex); handled += handle_list(hd, &list); } } if (handled) decr_entries(handled); } static void process_trace_bufs(void) { while (wait_empty_entries()) __process_trace_bufs(); } static void clean_trace_bufs(void) { /* * No mutex needed here: we're only reading from the lists, * tracers are done */ while (dp_entries) __process_trace_bufs(); } static inline void read_err(int cpu, char *ifn) { if (errno != EAGAIN) fprintf(stderr, "Thread %d failed read of %s: %d/%s\n", cpu, ifn, errno, strerror(errno)); } static int net_sendfile(struct io_info *iop) { int ret; ret = sendfile(iop->ofd, iop->ifd, NULL, iop->ready); if (ret < 0) { perror("sendfile"); return 1; } else if (ret < (int)iop->ready) { fprintf(stderr, "short sendfile send (%d of %d)\n", ret, iop->ready); return 1; } return 0; } static inline int net_sendfile_data(struct tracer *tp, struct io_info *iop) { struct devpath *dpp = iop->dpp; if (net_send_header(iop->ofd, tp->cpu, dpp->buts_name, iop->ready)) return 1; return net_sendfile(iop); } static int fill_ofname(char *dst, int dstlen, char *subdir, char *buts_name, int cpu) { int len; struct stat sb; if (output_dir) len = snprintf(dst, dstlen, "%s/", output_dir); else len = snprintf(dst, dstlen, "./"); if (subdir) len += snprintf(dst + len, dstlen - len, "%s", subdir); if (stat(dst, &sb) < 0) { if (errno != ENOENT) { fprintf(stderr, "Destination dir %s stat failed: %d/%s\n", dst, errno, strerror(errno)); return 1; } /* * There is no synchronization between multiple threads * trying to create the directory at once. It's harmless * to let them try, so just detect the problem and move on. */ if (mkdir(dst, 0755) < 0 && errno != EEXIST) { fprintf(stderr, "Destination dir %s can't be made: %d/%s\n", dst, errno, strerror(errno)); return 1; } } if (output_name) snprintf(dst + len, dstlen - len, "%s.blktrace.%d", output_name, cpu); else snprintf(dst + len, dstlen - len, "%s.blktrace.%d", buts_name, cpu); return 0; } static int set_vbuf(struct io_info *iop, int mode, size_t size) { iop->obuf = malloc(size); if (setvbuf(iop->ofp, iop->obuf, mode, size) < 0) { fprintf(stderr, "setvbuf(%s, %d) failed: %d/%s\n", iop->dpp->path, (int)size, errno, strerror(errno)); free(iop->obuf); return 1; } return 0; } static int iop_open(struct io_info *iop, int cpu) { char hostdir[MAXPATHLEN + 64]; iop->ofd = -1; if (net_mode == Net_server) { struct cl_conn *nc = iop->nc; int len; len = snprintf(hostdir, sizeof(hostdir), "%s-", nc->ch->hostname); len += strftime(hostdir + len, sizeof(hostdir) - len, "%F-%T/", gmtime(&iop->dpp->cl_connect_time)); } else { hostdir[0] = 0; } if (fill_ofname(iop->ofn, sizeof(iop->ofn), hostdir, iop->dpp->buts_name, cpu)) return 1; iop->ofp = my_fopen(iop->ofn, "w+"); if (iop->ofp == NULL) { fprintf(stderr, "Open output file %s failed: %d/%s\n", iop->ofn, errno, strerror(errno)); return 1; } if (set_vbuf(iop, _IOLBF, FILE_VBUF_SIZE)) { fprintf(stderr, "set_vbuf for file %s failed: %d/%s\n", iop->ofn, errno, strerror(errno)); fclose(iop->ofp); return 1; } iop->ofd = fileno(iop->ofp); return 0; } static void close_iop(struct io_info *iop) { struct mmap_info *mip = &iop->mmap_info; if (mip->fs_buf) munmap(mip->fs_buf, mip->fs_buf_len); if (!piped_output) { if (ftruncate(fileno(iop->ofp), mip->fs_size) < 0) { fprintf(stderr, "Ignoring err: ftruncate(%s): %d/%s\n", iop->ofn, errno, strerror(errno)); } } if (iop->ofp) fclose(iop->ofp); if (iop->obuf) free(iop->obuf); } static void close_ios(struct tracer *tp) { while (tp->nios > 0) { struct io_info *iop = &tp->ios[--tp->nios]; iop->dpp->drops = get_drops(iop->dpp); if (iop->ifd >= 0) close(iop->ifd); if (iop->ofp) close_iop(iop); else if (iop->ofd >= 0) { struct devpath *dpp = iop->dpp; net_send_close(iop->ofd, dpp->buts_name, dpp->drops); net_close_connection(&iop->ofd); } } free(tp->ios); free(tp->pfds); } static int open_ios(struct tracer *tp) { struct pollfd *pfd; struct io_info *iop; struct list_head *p; tp->ios = calloc(ndevs, sizeof(struct io_info)); memset(tp->ios, 0, ndevs * sizeof(struct io_info)); tp->pfds = calloc(ndevs, sizeof(struct pollfd)); memset(tp->pfds, 0, ndevs * sizeof(struct pollfd)); tp->nios = 0; iop = tp->ios; pfd = tp->pfds; __list_for_each(p, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); iop->dpp = dpp; iop->ofd = -1; snprintf(iop->ifn, sizeof(iop->ifn), "%s/block/%s/trace%d", debugfs_path, dpp->buts_name, tp->cpu); iop->ifd = my_open(iop->ifn, O_RDONLY | O_NONBLOCK); if (iop->ifd < 0) { fprintf(stderr, "Thread %d failed open %s: %d/%s\n", tp->cpu, iop->ifn, errno, strerror(errno)); return 1; } init_mmap_info(&iop->mmap_info); pfd->fd = iop->ifd; pfd->events = POLLIN; if (piped_output) ; else if (net_client_use_sendfile()) { iop->ofd = net_setup_client(); if (iop->ofd < 0) goto err; net_send_open(iop->ofd, tp->cpu, dpp->buts_name); } else if (net_mode == Net_none) { if (iop_open(iop, tp->cpu)) goto err; } else { /* * This ensures that the server knows about all * connections & devices before _any_ closes */ net_send_open(cl_fds[tp->cpu], tp->cpu, dpp->buts_name); } pfd++; iop++; tp->nios++; } return 0; err: close(iop->ifd); /* tp->nios _not_ bumped */ close_ios(tp); return 1; } static int handle_pfds_file(struct tracer *tp, int nevs, int force_read) { struct mmap_info *mip; int i, ret, nentries = 0; struct pollfd *pfd = tp->pfds; struct io_info *iop = tp->ios; for (i = 0; nevs > 0 && i < ndevs; i++, pfd++, iop++) { if (pfd->revents & POLLIN || force_read) { mip = &iop->mmap_info; ret = setup_mmap(iop->ofd, buf_size, mip, tp); if (ret < 0) { pfd->events = 0; break; } ret = read(iop->ifd, mip->fs_buf + mip->fs_off, buf_size); if (ret > 0) { pdc_dr_update(iop->dpp, tp->cpu, ret); mip->fs_size += ret; mip->fs_off += ret; nentries++; } else if (ret == 0) { /* * Short reads after we're done stop us * from trying reads. */ if (tp->is_done) clear_events(pfd); } else { read_err(tp->cpu, iop->ifn); if (errno != EAGAIN || tp->is_done) clear_events(pfd); } nevs--; } } return nentries; } static int handle_pfds_netclient(struct tracer *tp, int nevs, int force_read) { struct stat sb; int i, nentries = 0; struct pollfd *pfd = tp->pfds; struct io_info *iop = tp->ios; for (i = 0; i < ndevs; i++, pfd++, iop++) { if (pfd->revents & POLLIN || force_read) { if (fstat(iop->ifd, &sb) < 0) { perror(iop->ifn); pfd->events = 0; } else if (sb.st_size > (off_t)iop->data_queued) { iop->ready = sb.st_size - iop->data_queued; iop->data_queued = sb.st_size; if (!net_sendfile_data(tp, iop)) { pdc_dr_update(iop->dpp, tp->cpu, iop->ready); nentries++; } else clear_events(pfd); } if (--nevs == 0) break; } } if (nentries) incr_entries(nentries); return nentries; } static int handle_pfds_entries(struct tracer *tp, int nevs, int force_read) { int i, nentries = 0; struct trace_buf *tbp; struct pollfd *pfd = tp->pfds; struct io_info *iop = tp->ios; tbp = alloc_trace_buf(tp->cpu, buf_size); for (i = 0; i < ndevs; i++, pfd++, iop++) { if (pfd->revents & POLLIN || force_read) { tbp->len = read(iop->ifd, tbp->buf, buf_size); if (tbp->len > 0) { pdc_dr_update(iop->dpp, tp->cpu, tbp->len); add_trace_buf(iop->dpp, tp->cpu, &tbp); nentries++; } else if (tbp->len == 0) { /* * Short reads after we're done stop us * from trying reads. */ if (tp->is_done) clear_events(pfd); } else { read_err(tp->cpu, iop->ifn); if (errno != EAGAIN || tp->is_done) clear_events(pfd); } if (!piped_output && --nevs == 0) break; } } free(tbp); if (nentries) incr_entries(nentries); return nentries; } static void *thread_main(void *arg) { int ret, ndone, to_val; struct tracer *tp = arg; ret = lock_on_cpu(tp->cpu); if (ret) goto err; ret = open_ios(tp); if (ret) goto err; if (piped_output) to_val = 50; /* Frequent partial handles */ else to_val = 500; /* 1/2 second intervals */ tracer_signal_ready(tp, Th_running, 0); tracer_wait_unblock(tp); while (!tp->is_done) { ndone = poll(tp->pfds, ndevs, to_val); if (ndone || piped_output) (void)handle_pfds(tp, ndone, piped_output); else if (ndone < 0 && errno != EINTR) fprintf(stderr, "Thread %d poll failed: %d/%s\n", tp->cpu, errno, strerror(errno)); } /* * Trace is stopped, pull data until we get a short read */ while (handle_pfds(tp, ndevs, 1) > 0) ; close_ios(tp); tracer_signal_ready(tp, Th_leaving, 0); return NULL; err: tracer_signal_ready(tp, Th_error, ret); return NULL; } static int start_tracer(int cpu) { struct tracer *tp; tp = malloc(sizeof(*tp)); memset(tp, 0, sizeof(*tp)); INIT_LIST_HEAD(&tp->head); tp->status = 0; tp->cpu = cpu; if (pthread_create(&tp->thread, NULL, thread_main, tp)) { fprintf(stderr, "FAILED to start thread on CPU %d: %d/%s\n", cpu, errno, strerror(errno)); free(tp); return 1; } list_add_tail(&tp->head, &tracers); return 0; } static int create_output_files(int cpu) { char fname[MAXPATHLEN + 64]; struct list_head *p; FILE *f; __list_for_each(p, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); if (fill_ofname(fname, sizeof(fname), NULL, dpp->buts_name, cpu)) return 1; f = my_fopen(fname, "w+"); if (!f) return 1; fclose(f); } return 0; } static void start_tracers(void) { int cpu, started = 0; struct list_head *p; size_t alloc_size = CPU_ALLOC_SIZE(max_cpus); for (cpu = 0; cpu < max_cpus; cpu++) { if (!CPU_ISSET_S(cpu, alloc_size, online_cpus)) { /* * Create fake empty output files so that other tools * like blkparse don't have to bother with sparse CPU * number space. */ if (create_output_files(cpu)) break; continue; } if (start_tracer(cpu)) break; started++; } wait_tracers_ready(started); __list_for_each(p, &tracers) { struct tracer *tp = list_entry(p, struct tracer, head); if (tp->status) fprintf(stderr, "FAILED to start thread on CPU %d: %d/%s\n", tp->cpu, tp->status, strerror(tp->status)); } } static void stop_tracers(void) { struct list_head *p; /* * Stop the tracing - makes the tracer threads clean up quicker. */ __list_for_each(p, &devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); (void)ioctl(dpp->fd, BLKTRACESTOP); } /* * Tell each tracer to quit */ __list_for_each(p, &tracers) { struct tracer *tp = list_entry(p, struct tracer, head); tp->is_done = 1; } pthread_cond_broadcast(&mt_cond); } static void del_tracers(void) { struct list_head *p, *q; list_for_each_safe(p, q, &tracers) { struct tracer *tp = list_entry(p, struct tracer, head); list_del(&tp->head); free(tp); } } static void wait_tracers(void) { struct list_head *p; if (use_tracer_devpaths()) process_trace_bufs(); wait_tracers_leaving(); __list_for_each(p, &tracers) { int ret; struct tracer *tp = list_entry(p, struct tracer, head); ret = pthread_join(tp->thread, NULL); if (ret) fprintf(stderr, "Thread join %d failed %d\n", tp->cpu, ret); } if (use_tracer_devpaths()) clean_trace_bufs(); get_all_drops(); } static void exit_tracing(void) { signal(SIGINT, SIG_IGN); signal(SIGHUP, SIG_IGN); signal(SIGTERM, SIG_IGN); signal(SIGALRM, SIG_IGN); stop_tracers(); wait_tracers(); del_tracers(); rel_devpaths(); } static void handle_sigint(__attribute__((__unused__)) int sig) { done = 1; stop_tracers(); } static void show_stats(struct list_head *devpaths) { FILE *ofp; struct list_head *p; unsigned long long nevents, data_read; unsigned long long total_drops = 0; unsigned long long total_events = 0; if (piped_output) ofp = my_fopen("/dev/null", "w"); else ofp = stdout; __list_for_each(p, devpaths) { int cpu; struct pdc_stats *sp; struct devpath *dpp = list_entry(p, struct devpath, head); if (net_mode == Net_server) printf("server: end of run for %s:%s\n", dpp->ch->hostname, dpp->buts_name); data_read = 0; nevents = 0; fprintf(ofp, "=== %s ===\n", dpp->buts_name); for (cpu = 0, sp = dpp->stats; cpu < dpp->ncpus; cpu++, sp++) { /* * Estimate events if not known... */ if (sp->nevents == 0) { sp->nevents = sp->data_read / sizeof(struct blk_io_trace); } fprintf(ofp, " CPU%3d: %20llu events, %8llu KiB data\n", cpu, sp->nevents, (sp->data_read + 1023) >> 10); data_read += sp->data_read; nevents += sp->nevents; } fprintf(ofp, " Total: %20llu events (dropped %llu)," " %8llu KiB data\n", nevents, dpp->drops, (data_read + 1024) >> 10); total_drops += dpp->drops; total_events += (nevents + dpp->drops); } fflush(ofp); if (piped_output) fclose(ofp); if (total_drops) { double drops_ratio = 1.0; if (total_events) drops_ratio = (double)total_drops/(double)total_events; fprintf(stderr, "\nYou have %llu (%5.1lf%%) dropped events\n" "Consider using a larger buffer size (-b) " "and/or more buffers (-n)\n", total_drops, 100.0 * drops_ratio); } } static int handle_args(int argc, char *argv[]) { int c, i; struct statfs st; int act_mask_tmp = 0; while ((c = getopt_long(argc, argv, S_OPTS, l_opts, NULL)) >= 0) { switch (c) { case 'a': i = find_mask_map(optarg); if (i < 0) { fprintf(stderr, "Invalid action mask %s\n", optarg); return 1; } act_mask_tmp |= i; break; case 'A': if ((sscanf(optarg, "%x", &i) != 1) || !valid_act_opt(i)) { fprintf(stderr, "Invalid set action mask %s/0x%x\n", optarg, i); return 1; } act_mask_tmp = i; break; case 'd': if (add_devpath(optarg) != 0) return 1; break; case 'I': { char dev_line[256]; FILE *ifp = my_fopen(optarg, "r"); if (!ifp) { fprintf(stderr, "Invalid file for devices %s\n", optarg); return 1; } while (fscanf(ifp, "%s\n", dev_line) == 1) { if (add_devpath(dev_line) != 0) { fclose(ifp); return 1; } } fclose(ifp); break; } case 'r': debugfs_path = optarg; break; case 'o': output_name = optarg; break; case 'k': kill_running_trace = 1; break; case 'w': stop_watch = atoi(optarg); if (stop_watch <= 0) { fprintf(stderr, "Invalid stopwatch value (%d secs)\n", stop_watch); return 1; } break; case 'V': case 'v': printf("%s version %s\n", argv[0], blktrace_version); exit(0); /*NOTREACHED*/ case 'b': buf_size = strtoul(optarg, NULL, 10); if (buf_size <= 0 || buf_size > 16*1024) { fprintf(stderr, "Invalid buffer size (%lu)\n", buf_size); return 1; } buf_size <<= 10; break; case 'n': buf_nr = strtoul(optarg, NULL, 10); if (buf_nr <= 0) { fprintf(stderr, "Invalid buffer nr (%lu)\n", buf_nr); return 1; } break; case 'D': output_dir = optarg; break; case 'h': net_mode = Net_client; memset(hostname, 0, sizeof(hostname)); strncpy(hostname, optarg, sizeof(hostname)); hostname[sizeof(hostname) - 1] = '\0'; break; case 'l': net_mode = Net_server; break; case 'p': net_port = atoi(optarg); break; case 's': net_use_sendfile = 0; break; default: show_usage(argv[0]); exit(1); /*NOTREACHED*/ } } while (optind < argc) if (add_devpath(argv[optind++]) != 0) return 1; if (net_mode != Net_server && ndevs == 0) { show_usage(argv[0]); return 1; } if (statfs(debugfs_path, &st) < 0) { fprintf(stderr, "Invalid debug path %s: %d/%s\n", debugfs_path, errno, strerror(errno)); return 1; } if (st.f_type != (long)DEBUGFS_TYPE) { fprintf(stderr, "Debugfs is not mounted at %s\n", debugfs_path); return 1; } if (act_mask_tmp != 0) act_mask = act_mask_tmp; if (net_mode == Net_client && net_setup_addr()) return 1; /* * Set up for appropriate PFD handler based upon output name. */ if (net_client_use_sendfile()) handle_pfds = handle_pfds_netclient; else if (net_client_use_send()) handle_pfds = handle_pfds_entries; else if (output_name && (strcmp(output_name, "-") == 0)) { piped_output = 1; handle_pfds = handle_pfds_entries; pfp = stdout; if (setvbuf(pfp, NULL, _IONBF, 0)) { perror("setvbuf stdout"); return 1; } } else handle_pfds = handle_pfds_file; return 0; } static void ch_add_connection(struct net_server_s *ns, struct cl_host *ch, int fd) { struct cl_conn *nc; nc = malloc(sizeof(*nc)); memset(nc, 0, sizeof(*nc)); time(&nc->connect_time); nc->ch = ch; nc->fd = fd; nc->ncpus = -1; list_add_tail(&nc->ch_head, &ch->conn_list); ch->connects++; list_add_tail(&nc->ns_head, &ns->conn_list); ns->connects++; ns->pfds = realloc(ns->pfds, (ns->connects+1) * sizeof(struct pollfd)); } static void ch_rem_connection(struct net_server_s *ns, struct cl_host *ch, struct cl_conn *nc) { net_close_connection(&nc->fd); list_del(&nc->ch_head); ch->connects--; list_del(&nc->ns_head); ns->connects--; ns->pfds = realloc(ns->pfds, (ns->connects+1) * sizeof(struct pollfd)); free(nc); } static struct cl_host *net_find_client_host(struct net_server_s *ns, struct in_addr cl_in_addr) { struct list_head *p; __list_for_each(p, &ns->ch_list) { struct cl_host *ch = list_entry(p, struct cl_host, head); if (in_addr_eq(ch->cl_in_addr, cl_in_addr)) return ch; } return NULL; } static struct cl_host *net_add_client_host(struct net_server_s *ns, struct sockaddr_in *addr) { struct cl_host *ch; ch = malloc(sizeof(*ch)); memset(ch, 0, sizeof(*ch)); ch->ns = ns; ch->cl_in_addr = addr->sin_addr; list_add_tail(&ch->head, &ns->ch_list); ns->nchs++; ch->hostname = strdup(inet_ntoa(addr->sin_addr)); printf("server: connection from %s\n", ch->hostname); INIT_LIST_HEAD(&ch->conn_list); INIT_LIST_HEAD(&ch->devpaths); return ch; } static void device_done(struct devpath *dpp, int ncpus) { int cpu; struct io_info *iop; for (cpu = 0, iop = dpp->ios; cpu < ncpus; cpu++, iop++) close_iop(iop); list_del(&dpp->head); dpp_free(dpp); } static void net_ch_remove(struct cl_host *ch, int ncpus) { struct list_head *p, *q; struct net_server_s *ns = ch->ns; list_for_each_safe(p, q, &ch->devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); device_done(dpp, ncpus); } list_for_each_safe(p, q, &ch->conn_list) { struct cl_conn *nc = list_entry(p, struct cl_conn, ch_head); ch_rem_connection(ns, ch, nc); } list_del(&ch->head); ns->nchs--; if (ch->hostname) free(ch->hostname); free(ch); } static void net_add_connection(struct net_server_s *ns) { int fd; struct cl_host *ch; socklen_t socklen = sizeof(ns->addr); fd = my_accept(ns->listen_fd, (struct sockaddr *)&ns->addr, &socklen); if (fd < 0) { /* * This is OK: we just won't accept this connection, * nothing fatal. */ perror("accept"); } else { ch = net_find_client_host(ns, ns->addr.sin_addr); if (!ch) ch = net_add_client_host(ns, &ns->addr); ch_add_connection(ns, ch, fd); } } static struct devpath *nc_add_dpp(struct cl_conn *nc, struct blktrace_net_hdr *bnh, time_t connect_time) { int cpu; struct io_info *iop; struct devpath *dpp; dpp = malloc(sizeof(*dpp)); memset(dpp, 0, sizeof(*dpp)); dpp->buts_name = strdup(bnh->buts_name); dpp->path = strdup(bnh->buts_name); dpp->fd = -1; dpp->ch = nc->ch; dpp->cl_id = bnh->cl_id; dpp->cl_connect_time = connect_time; dpp->ncpus = nc->ncpus; dpp->stats = calloc(dpp->ncpus, sizeof(*dpp->stats)); memset(dpp->stats, 0, dpp->ncpus * sizeof(*dpp->stats)); list_add_tail(&dpp->head, &nc->ch->devpaths); nc->ch->ndevs++; dpp->ios = calloc(nc->ncpus, sizeof(*iop)); memset(dpp->ios, 0, ndevs * sizeof(*iop)); for (cpu = 0, iop = dpp->ios; cpu < nc->ncpus; cpu++, iop++) { iop->dpp = dpp; iop->nc = nc; init_mmap_info(&iop->mmap_info); if (iop_open(iop, cpu)) goto err; } return dpp; err: /* * Need to unravel what's been done... */ while (cpu >= 0) close_iop(&dpp->ios[cpu--]); dpp_free(dpp); return NULL; } static struct devpath *nc_find_dpp(struct cl_conn *nc, struct blktrace_net_hdr *bnh) { struct list_head *p; time_t connect_time = nc->connect_time; __list_for_each(p, &nc->ch->devpaths) { struct devpath *dpp = list_entry(p, struct devpath, head); if (!strcmp(dpp->buts_name, bnh->buts_name)) return dpp; if (dpp->cl_id == bnh->cl_id) connect_time = dpp->cl_connect_time; } return nc_add_dpp(nc, bnh, connect_time); } static void net_client_read_data(struct cl_conn *nc, struct devpath *dpp, struct blktrace_net_hdr *bnh) { int ret; struct io_info *iop = &dpp->ios[bnh->cpu]; struct mmap_info *mip = &iop->mmap_info; if (setup_mmap(iop->ofd, bnh->len, &iop->mmap_info, NULL)) { fprintf(stderr, "ncd(%s:%d): mmap failed\n", nc->ch->hostname, nc->fd); exit(1); } ret = net_recv_data(nc->fd, mip->fs_buf + mip->fs_off, bnh->len); if (ret > 0) { pdc_dr_update(dpp, bnh->cpu, ret); mip->fs_size += ret; mip->fs_off += ret; } else if (ret < 0) exit(1); } /* * Returns 1 if we closed a host - invalidates other polling information * that may be present. */ static int net_client_data(struct cl_conn *nc) { int ret; struct devpath *dpp; struct blktrace_net_hdr bnh; ret = net_get_header(nc, &bnh); if (ret == 0) return 0; if (ret < 0) { fprintf(stderr, "ncd(%d): header read failed\n", nc->fd); exit(1); } if (data_is_native == -1 && check_data_endianness(bnh.magic)) { fprintf(stderr, "ncd(%d): received data is bad\n", nc->fd); exit(1); } if (!data_is_native) { bnh.magic = be32_to_cpu(bnh.magic); bnh.cpu = be32_to_cpu(bnh.cpu); bnh.max_cpus = be32_to_cpu(bnh.max_cpus); bnh.len = be32_to_cpu(bnh.len); bnh.cl_id = be32_to_cpu(bnh.cl_id); bnh.buf_size = be32_to_cpu(bnh.buf_size); bnh.buf_nr = be32_to_cpu(bnh.buf_nr); bnh.page_size = be32_to_cpu(bnh.page_size); } if ((bnh.magic & 0xffffff00) != BLK_IO_TRACE_MAGIC) { fprintf(stderr, "ncd(%s:%d): bad data magic\n", nc->ch->hostname, nc->fd); exit(1); } if (nc->ncpus == -1) nc->ncpus = bnh.max_cpus; /* * len == 0 means the other end is sending us a new connection/dpp * len == 1 means that the other end signalled end-of-run */ dpp = nc_find_dpp(nc, &bnh); if (bnh.len == 0) { /* * Just adding in the dpp above is enough */ ack_open_close(nc->fd, dpp->buts_name); nc->ch->cl_opens++; } else if (bnh.len == 1) { /* * overload cpu count with dropped events */ dpp->drops = bnh.cpu; ack_open_close(nc->fd, dpp->buts_name); if (--nc->ch->cl_opens == 0) { show_stats(&nc->ch->devpaths); net_ch_remove(nc->ch, nc->ncpus); return 1; } } else net_client_read_data(nc, dpp, &bnh); return 0; } static void handle_client_data(struct net_server_s *ns, int events) { struct cl_conn *nc; struct pollfd *pfd; struct list_head *p, *q; pfd = &ns->pfds[1]; list_for_each_safe(p, q, &ns->conn_list) { if (pfd->revents & POLLIN) { nc = list_entry(p, struct cl_conn, ns_head); if (net_client_data(nc) || --events == 0) break; } pfd++; } } static void net_setup_pfds(struct net_server_s *ns) { struct pollfd *pfd; struct list_head *p; ns->pfds[0].fd = ns->listen_fd; ns->pfds[0].events = POLLIN; pfd = &ns->pfds[1]; __list_for_each(p, &ns->conn_list) { struct cl_conn *nc = list_entry(p, struct cl_conn, ns_head); pfd->fd = nc->fd; pfd->events = POLLIN; pfd++; } } static int net_server_handle_connections(struct net_server_s *ns) { int events; printf("server: waiting for connections...\n"); while (!done) { net_setup_pfds(ns); events = poll(ns->pfds, ns->connects + 1, -1); if (events < 0) { if (errno != EINTR) { perror("FATAL: poll error"); return 1; } } else if (events > 0) { if (ns->pfds[0].revents & POLLIN) { net_add_connection(ns); events--; } if (events) handle_client_data(ns, events); } } return 0; } static int net_server(void) { int fd, opt; int ret = 1; struct net_server_s net_server; struct net_server_s *ns = &net_server; memset(ns, 0, sizeof(*ns)); INIT_LIST_HEAD(&ns->ch_list); INIT_LIST_HEAD(&ns->conn_list); ns->pfds = malloc(sizeof(struct pollfd)); fd = my_socket(AF_INET, SOCK_STREAM, 0); if (fd < 0) { perror("server: socket"); goto out; } opt = 1; if (setsockopt(fd, SOL_SOCKET, SO_REUSEADDR, &opt, sizeof(opt)) < 0) { perror("setsockopt"); goto out; } memset(&ns->addr, 0, sizeof(ns->addr)); ns->addr.sin_family = AF_INET; ns->addr.sin_addr.s_addr = htonl(INADDR_ANY); ns->addr.sin_port = htons(net_port); if (bind(fd, (struct sockaddr *) &ns->addr, sizeof(ns->addr)) < 0) { perror("bind"); goto out; } if (listen(fd, 1) < 0) { perror("listen"); goto out; } /* * The actual server looping is done here: */ ns->listen_fd = fd; ret = net_server_handle_connections(ns); /* * Clean up and return... */ out: free(ns->pfds); return ret; } static int run_tracers(void) { atexit(exit_tracing); if (net_mode == Net_client) printf("blktrace: connecting to %s\n", hostname); if (setup_buts()) return 1; if (use_tracer_devpaths()) { if (setup_tracer_devpaths()) return 1; if (piped_output) handle_list = handle_list_file; else handle_list = handle_list_net; } start_tracers(); if (nthreads_running == ncpus) { unblock_tracers(); start_buts(); if (net_mode == Net_client) printf("blktrace: connected!\n"); if (stop_watch) alarm(stop_watch); } else stop_tracers(); wait_tracers(); if (nthreads_running == ncpus) show_stats(&devpaths); if (net_client_use_send()) close_client_connections(); del_tracers(); return 0; } static cpu_set_t *get_online_cpus(void) { FILE *cpus; cpu_set_t *set; size_t alloc_size; int cpuid, prevcpuid = -1; char nextch; int n, ncpu, curcpu = 0; int *cpu_nums; ncpu = sysconf(_SC_NPROCESSORS_CONF); if (ncpu < 0) return NULL; cpu_nums = malloc(sizeof(int)*ncpu); if (!cpu_nums) { errno = ENOMEM; return NULL; } /* * There is no way to easily get maximum CPU number. So we have to * parse the file first to find it out and then create appropriate * cpuset */ cpus = my_fopen("/sys/devices/system/cpu/online", "r"); for (;;) { n = fscanf(cpus, "%d%c", &cpuid, &nextch); if (n <= 0) break; if (n == 2 && nextch == '-') { prevcpuid = cpuid; continue; } if (prevcpuid == -1) prevcpuid = cpuid; while (prevcpuid <= cpuid) { /* More CPUs listed than configured? */ if (curcpu >= ncpu) { errno = EINVAL; return NULL; } cpu_nums[curcpu++] = prevcpuid++; } prevcpuid = -1; } fclose(cpus); ncpu = curcpu; max_cpus = cpu_nums[ncpu - 1] + 1; /* Now that we have maximum cpu number, create a cpuset */ set = CPU_ALLOC(max_cpus); if (!set) { errno = ENOMEM; return NULL; } alloc_size = CPU_ALLOC_SIZE(max_cpus); CPU_ZERO_S(alloc_size, set); for (curcpu = 0; curcpu < ncpu; curcpu++) CPU_SET_S(cpu_nums[curcpu], alloc_size, set); free(cpu_nums); return set; } int main(int argc, char *argv[]) { int ret = 0; setlocale(LC_NUMERIC, "en_US"); pagesize = getpagesize(); online_cpus = get_online_cpus(); if (!online_cpus) { fprintf(stderr, "cannot get online cpus %d/%s\n", errno, strerror(errno)); ret = 1; goto out; } else if (handle_args(argc, argv)) { ret = 1; goto out; } ncpus = CPU_COUNT_S(CPU_ALLOC_SIZE(max_cpus), online_cpus); if (ndevs > 1 && output_name && strcmp(output_name, "-") != 0) { fprintf(stderr, "-o not supported with multiple devices\n"); ret = 1; goto out; } signal(SIGINT, handle_sigint); signal(SIGHUP, handle_sigint); signal(SIGTERM, handle_sigint); signal(SIGALRM, handle_sigint); signal(SIGPIPE, SIG_IGN); if (kill_running_trace) { struct devpath *dpp; struct list_head *p; __list_for_each(p, &devpaths) { dpp = list_entry(p, struct devpath, head); if (__stop_trace(dpp->fd)) { fprintf(stderr, "BLKTRACETEARDOWN %s failed: %d/%s\n", dpp->path, errno, strerror(errno)); } } } else if (net_mode == Net_server) { if (output_name) { fprintf(stderr, "-o ignored in server mode\n"); output_name = NULL; } ret = net_server(); } else ret = run_tracers(); out: if (pfp) fclose(pfp); rel_devpaths(); return ret; }