Computer Communication Review: Papers

Find a CCR issue:
  • Sean Donovan, Nick Feamster

    Home and business network operators have limited network statistics available over which management decisions can be made. Similarly, there are few triggered behaviors, such as usage or bandwidths cap for individual users, that are available. By looking at sources of traffic, based on Domain Name System (DNS) cues for content of particular web addresses or source Autonomous System (AS) of the traffic, network operators could create new and interesting rules for their network. NetAssay is a Software-Defined Networking (SDN)-based, network-wide monitoring and reaction framework. By integrating information from Border Gateway Protocol (BGP) and the Domain Name System, NetAssay is able to integrate formerly disparate sources of control information, and use it to provide better monitoring, more useful triggered events, and security benefits for network operators.

  • Maksym Gabielkov, Ashwin Rao, Arnaud Legout

    Online social networks (OSNs) are an important source of information for scientists in different fields such as computer science, sociology, economics, etc. However, it is hard to study OSNs as they are very large. For instance, Facebook has 1.28 billion active users in March 2014 and Twitter claims 255 million active users in April 2014. Also, companies take measures to prevent crawls of their OSNs and refrain from sharing their data with the research community. For these reasons, we argue that sampling techniques will be the best technique to study OSNs in the future. In this work, we take an experimental approach to study the characteristics of well-known sampling techniques on a full social graph of Twitter crawled in 2012 [2]. Our contribution is to evaluate the behavior of these techniques on a real directed graph by considering two sampling scenarios: (a) obtaining most popular users (b) obtaining an unbiased sample of users, and to find the most suitable sampling techniques for each scenario.

  • Benjamin Hesmans, Olivier Bonaventure
  • Jinzhen Bao, Baokang Zhao, Wanrong Yu, Zhenqian Feng, Chunqing Wu, Zhenghu Gong

    In recent years, with the rapid development of satellite technology including On Board Processing (OBP) and Inter Satellite Link (ISL), satellite network devices such as space IP routers have been experimentally carried in space. However, there are many difficulties to build a future satellite network with current terrestrial Internet technologies due to the distinguished space features, such as the severely limited resources, remote hardware/software upgrade in space. In this paper, we propose OpenSAN, a novel architecture of software-defined satellite network. By decoupling the data plane and control plane, OpenSAN provides satellite network with high efficiency, finegrained control, as well as flexibility to support future advanced network technology. Furthermore, we also discuss some practical challenges in the deployment of OpenSAN.

  • Ravi Netravali, Anirudh Sivaraman, Keith Winstein, Somak Das, Ameesh Goyal, Hari Balakrishnan

    This demo presents a measurement toolkit, Mahimahi, that records websites and replays them under emulated network conditions. Mahimahi is structured as a set of arbitrarily composable UNIX shells. It includes two shells to record and replay Web pages, RecordShell and ReplayShell, as well as two shells for network emulation, DelayShell and LinkShell. In addition, Mahimahi includes a corpus of recorded websites along with benchmark results and link traces (https://github.com/ravinet/sites). Mahimahi improves on prior record-and-replay frameworks in three ways. First, it preserves the multi-origin nature of Web pages, present in approximately 98% of the Alexa U.S. Top 500, when replaying. Second, Mahimahi isolates its own network traffic, allowing multiple instances to run concurrently with no impact on the host machine and collected measurements. Finally, Mahimahi is not inherently tied to browsers and can be used to evaluate many different applications. A demo of Mahimahi recording and replaying a Web page over an emulated link can be found at http://youtu.be/vytwDKBA-8s. The source code and instructions to use Mahimahi are available at http://mahimahi.mit.edu/.

  • Mo Dong, Qingxi Li, Doron Zarchy, Brighten Godfrey, Michael Schapira

    After more than two decades of evolution, TCP and its end host based modifications can still suffer from severely degraded performance under real-world challenging network conditions. The reason, as we observe, is due to TCP family’s fundamental architectural deficiency, which hardwires packet-level events to control responses and ignores emprical performance. Jumping out of TCP lineage’s architectural deficiency, we propose Performanceoriented Congestion Control (PCC), a new congestion control architecture in which each sender controls its sending strategy based on empirically observed performance metrics. We show through preliminary experimental results that PCC achieves consistently high performance under various challenging network conditions.

  • Arup Raton Roy, Md. Faizul Bari, Mohamed Faten Zhani, Reaz Ahmed, Raouf Boutaba
  • Adrian Gämperli, Vasileios Kotronis, Xenofontas Dimitropoulos
  • Abdulla Alwabel, Minlan Yu, Ying Zhang, Jelena Mirkovic

    We propose a new software-defined security service – SENSS – that enables a victim network to request services from remote ISPs for traffic that carries source IPs or destination IPs from this network’s address space. These services range from statistics gathering, to filtering or quality of service guarantees, to route reports or modifications. The SENSS service has very simple, yet powerful, interfaces. This enables it to handle a variety of data plane and control plane attacks, while being easily implementable in today’s ISP. Through extensive evaluations on realistic traffic traces and Internet topology, we show how SENSS can be used to quickly, safely and effectively mitigate a variety of large-scale attacks that are largely unhandled today.

  • Zachary S. Bischof, Fabián E. Bustamante
  • Pierdomenico Fiadino, Mirko Schiavone, Pedro Casas

    WhatsApp, the new giant in instant multimedia messaging in mobile networks is rapidly increasing its popularity, taking over the traditional SMS/MMS messaging. In this paper we present the first large-scale characterization of WhatsApp, useful among others to ISPs willing to understand the impacts of this and similar applications on their networks. Through the combined analysis of passive measurements at the core of a national mobile network, worldwide geo-distributed active measurements, and tra c analysis at end devices, we show that: (i) the WhatsApp hosting architecture is highly centralized and exclusively located in the US; (ii) video sharing covers almost 40% of the total WhatsApp tra c volume; (iii) flow characteristics depend on the OS of the end device; (iv) despite the big latencies to US servers, download throughputs are as high as 1.5 Mbps; (v) users react immediately and negatively to service outages through social networks feedbacks.

  • Aisha Mushtaq, Asad Khalid Ismail, Abdul Wasay, Bilal Mahmood, Ihsan Ayyub Qazi, Zartash Afzal Uzmi

    Data center operators face extreme challenges in simultaneously providing low latency for short flows, high throughput for long flows, and high burst tolerance. We propose a buffer management strategy that addresses these challenges by isolating short and long flows into separate buffers, sizing these buffers based on flow requirements, and scheduling packets to meet different flow-level objectives. Our design provides new opportunities for performance improvements that complement transport layer optimizations.

  • Joel Obstfeld, Simon Knight, Ed Kern, Qiang Sheng Wang, Tom Bryan, Dan Bourque

    The increasing demand to provide new network services in a timely and efficient manner is driving the need to design, test and deploy networks quickly and consistently. Testing and verifying at scale is a challenge: network equipment is expensive, requires space, power and cooling, and there is never enough test equipment for everyone who wants to use it! Network virtualization technologies enable a flexible environment for educators, researchers, and operators to create functional models of current, planned, or theoretical networks. This demonstration will show VIRL — the Virtual Internet Routing Lab — a platform that can be used for network change validation, training, education, research, or networkaware applications development. The platform combines network virtualization technologies with virtual machines (VMs) running open-source and commercial operating systems; VM orchestration capabilities; a context-aware configuration engine; and an extensible data-collection framework. The system simplifies the process to create both simple and complex environments, run simulations, and collect measurement data.

  • Wentao Chang, An Wang, Aziz Mohaisen, Songqing Chen
  • John P. Rula, Fabian E. Bustamante
  • Shaofeng Chen, Dingyi Fang, Xiaojiang Chen, Tingting Xia, Meng Jin

    This poster presents GuideLoc, a highly efficient aerial wireless localization system that uses directional antennas mounted on a mini Multirotor Unmanned Aerial Vehicle (UAV) to enable detecting and positioning of targets. Taking advantage of angle and signal strength of frames transmitted from targets, GuideLoc can directly fly towards the targets with the minimum flight route and time. We implement a prototype of GuideLoc and evaluate the performance by simulations and experiments. Experimental results show that GuideLoc achieves an average location accuracy of 2.7 meters and reduces flight distance more than 50% compared with existing localization approaches using UAV.

  • Keunhong Lee, Joongi Kim, Sue Moon
  • Yuliang Li, Guang Yao, Jun Bi
  • Jun Li, Skyler Berg, Mingwei Zhang, Peter Reiher, Tao Wei

    End hosts in today’s Internet have the best knowledge of the type of traffic they should receive, but they play no active role in traffic engineering. Traffic engineering is conducted by ISPs, which unfortunately are blind to specific user needs. End hosts are therefore subject to unwanted traffic, particularly from Distributed Denial of Service (DDoS) attacks. This research proposes a new system called DrawBridge to address this traffic engineering dilemma. By realizing the potential of software-defined networking (SDN), in this research we investigate a solution that enables end hosts to use their knowledge of desired traffic to improve traffic engineering during DDoS attacks.

  • Angela H. Jiang, Zachary S. Bischof, Fabian E. Bustamante

    A social news site presents user-curated content, ranked by popularity. Popular curators like Reddit, or Facebook have become effective way of crowdsourcing news or sharing for personal opinions. Traditionally, these services require a centralized authority to aggregate data and determine what to display. However, the trust issues that arise from a centralized system are particularly damaging to the “Web democracy” that social news sites are meant to provide. In this poster, we present cliq, a decentralized social news curator. cliq is a P2P based social news curator that provides private and unbiased reporting. All users in cliq share responsibility for tracking and providing popular content. Any user data that cliq needs to store is also managed across the network. We first inform our design of cliq through an analysis of Reddit. We design a way to provide content curation without a persistent moderator, or usernames.

  • Matthias Vallentin, Dominik Charousset, Thomas C. Schmidt, Vern Paxson, Matthias Wählisch

    When an organization detects a security breach, it undertakes a forensic analysis to figure out what happened. This investigation involves inspecting a wide range of heterogeneous data sources spanning over a long period of time. The iterative nature of the analysis procedure requires an interactive experience with the data. However, the distributed processing paradigms we find in practice today fail to provide this requirement: the batch-oriented nature of MapReduce cannot deliver sub-second round-trip times, and distributed in-memory processing cannot store the terabytes of activity logs needed to inspect during an incident. We present the design and implementation of Visibility Across Space and Time (VAST), a distributed database to support interactive network forensics, and libcppa, its exceptionally scalable messaging core. The extended actor framework libcppa enables VAST to distribute lightweight tasks at negligible overhead. In our live demo, we showcase how VAST enables security analysts to grapple with the huge amounts of data often associated with incident investigations.

  • Chengchen Hu, Ji Yang, Zhimin Gong, Shuoling Deng, Hongbo Zhao
  • Arpit Gupta, Laurent Vanbever, Muhammad Shahbaz, Sean Patrick Donovan, Brandon Schlinker, Nick Feamster, Jennifer Rexford, Scott Shenker, Russ Clark, Ethan Katz-Bassett

    BGP severely constrains how networks can deliver traffic over the Internet. Today’s networks can only forward traffic based on the destination IP prefix, by selecting among routes offered by their immediate neighbors. We believe Software Defined Networking (SDN) could revolutionize wide-area traffic delivery, by offering direct control over packet-processing rules that match on multiple header fields and perform a variety of actions. Internet exchange points (IXPs) are a compelling place to start, given their central role in interconnecting many networks and their growing importance in bringing popular content closer to end users. To realize a Software Defined IXP (an “SDX”), need new programming abstractions that allow participating networks to create and run these applications and a runtime that both behaves correctly when interacting with BGP and ensures that applications do not interfere with each other. We must also ensure that the system scales, both in rule-table size and computational overhead. In this demo, we show how we tackle these challenges demonstrating the flexibility and scalability of our SDX platform. The paper also appears in the main program [1].

  • Dina Papagiannaki

    Welcome to the July issue of CCR, an issue that should hopefully inspire a number of discussions that we can continue in person during Sigcomm, in Chicago. This issue features 17 papers, 5 of which are editorial notes, and 12 technical contributions from our community. The technical part features novel contributions in the area of router location inference, performance of fiber-to-the-home networks, BGP, programmable middleboxes, and a programming language for protocol independent packet processors. Each one of them is advancing the state of the art and should be a useful building block for future research. The research community is increasingly becoming multidisciplinary. One cannot help but get inspired when he/she sees the elegance of solutions that address real problems in one discipline while exploiting knowledge produced in another. This is the mission of the fifth technical submission in this issue. The core of the contribution is to adopt the concept of design contests and apply it to the area of congestion control protocols in wireless networks. The authors point out that one of the key requirements in any design contest is to “have an unambigious, measurable objective that will allow one to compare protocols”. And this is exactly what the authors do in their work. The article concludes that design contests can benefit networking research, if designed properly, and they encourage others to explore their strengths and weaknesses. The remaining papers of the technical part are devoted to one of the largest efforts undertaken in the recent years to rethink the architecture of the Internet, e.g. the Future Internet Architecture program of the U.S. National Science Foundation. FIA targets the design of a trustworthy Internet, that incorporates societal, economical, and legal constraints, while following a clean slate approach. It was the initiative of Prof. David Wetherall, from University of Washington, to bring the four FIA proposals, and affiliated project ChoiceNet, to CCR, and provide a very comprehensive exposition of the different avenues taken by the different consortia. I have to thank David for acm all the hard work he did in order to bring all the pieces in the same place, something that will undoubtedly help our community understand the FIA efforts in a greater extent. The FIA session is preceded by a technical note by Dr. Darleen Fisher, FIA program director at the U.S. National Science Foundation. It is inspiring to see how a long term (11-year) funding effort has led to a number of functioning components that may define the Internet of the future. Thank you Darleen for a wonderful introductory note! Our editorial session comprises 5 papers. Two of the papers cover workshop reports: i) the workshop on Internet Economics 2013, and ii) the roundtable on real time communications research, that was held along with IPTComm, in October 2013. We have an article introducing the ProtoRINA prototype, a user-space prototype of the Recursive InterNetwork Architecture (RINA), and a qualitative study of the Internet census data that was collected in March 2013, and that has attracted significant attention in our community. The last editorial is appearing in CCR per my own invitation to its author, Daniel Stenberg. By the end of this year the Internet Engineering Task Force (IETF) is aiming to standardize the second version of HTTP, i.e. HTTP 2.0. This new version is going to be a very significant change compared to HTTP v1 aiming to provide better support for mobile browsing. Daniel is a Mozilla engineer participating in the standardization of HTTP 2.0 and has kindly accepted to publish his thoughts on HTTP 2.0 at CCR. This issue also marks the start of tenure for Dr. Aline Carneiro Viana, from INRIA. Aline is bringing a lot of energy to the editorial board and her expertise in ad hoc, sensor networks, delay tolerant networks, and cognitive radio networks. With all that, I hope to see most of you in Chicago in August, and please feel free to send me any suggestions on things you would like to see published from CCR in the future.

  • B. Huffaker, M. Fomenkov, K. Claffy

    In this paper we focus on geolocating Internet routers, using a methodology for extracting and decoding geography-related strings from fully qualified domain names (hostnames). We first compiled an extensive dictionary associating geographic strings (e.g., airport codes) with geophysical locations. We then searched a large set of router hostnames for these strings, assuming each autonomous naming domain uses geographic hints consistently within that domain. We used topology and performance data continually collected by our global measurement infrastructure to discern whether a given hint appears to co-locate different hostnames in which it is found. Finally, we generalized geolocation hints into domain-specific rule sets. We generated a total of 1,711 rules covering 1,398 different domains and validated them using domain-specific ground truth we gathered for six domains. Unlike previous efforts which relied on labor-intensive domain-specific manual analysis, we automate our process for inferring the domain specific heuristics, substantially advancing the state-of-the-art of methods for geolocating Internet resources.

    Joel Sommers
  • M. Luckie

    Researchers depend on public BGP data to understand the structure and evolution of the AS topology, as well as the operational security and resiliency of BGP. BGP data is provided voluntarily by network operators who establish BGP sessions with route collectors that record this data. In this paper, we show how trivial it is for a single vantage point (VP) to introduce thousands of spurious routes into the collection by providing examples of five VPs that did so. We explore the impact these misbehaving VPs had on AS relationship inference, showing these misbehaving VPs introduced thousands of AS links that did not exist, and caused relationship inferences for links that did exist to be corrupted. We evaluate methods to automatically identify misbehaving VPs, although we find the result unsatisfying because the limitations of real-world BGP practices and AS relationship inference algorithms produce signatures similar to those created by misbehaving VPs. The most recent misbehaving VP we discovered added thousands of spurious routes for nine consecutive months until 8 November 2012. This misbehaving VP barely impacts (0.1%) our validation of our AS relationship inferences, but this number may be misleading since most of our validation data relies on BGP and RPSL which validates only existing links, rather than asserting the non-existence of links. We have only a few assertions of non-existent routes, all received via our public-facing website that allows operators to provide validation data through our interactive feedback mechanism. We only discovered this misbehavior because two independent operators corrected some inferences, and we noticed that the spurious routes all came from the same VP. This event highlights the limitations of even the best available topology data, and provides additional evidence that comprehensive ground truth validation from operators is essential to scientific research on Internet topology.

    Renata Teixeira
  • M. Sargent, M. Allman

    Fiber-To-The-Home (FTTH) networks are on the brink of bringing significantly higher capacity to residential users compared to today'ss commercial residential options. There are several burgeoning FTTH networks that provide capacities of up to 1 Gbps. We have been monitoring one such operational network the Case Connection Zone - for 23 months. In this paper we seek to understand the extent to which the users in this network are in fact making use of the provided bi-directional 1 Gbps capacity. We find that even when given virtually unlimited capacity the majority of the time users do not retrieve information from the Internet in excess of commercially available data rates and transmit at only modestly higher rates than commodity networks support. Further, we find that end host issues - most prominently buffering at both end points - are often the cause of the lower-than-expected performance.

    Fabián E. Bustamante
  • K. Khan, Z. Ahmed, S. Ahmed, A. Syed, S. Khayam

    With only access billing no longer ensuring profits, an ISP's growth now relies on rolling out new and differentiated services. However, ISPs currently do not have a well-defined architecture for rapid, cost-effective, and scalable dissemination of new services. We present iSDF, a new SDN-enabled framework that can meet an ISP's service delivery constraints concerning cost, scalability, deployment flexibility, and operational ease. We show that meeting these constraints necessitates an SDN philosophy for a centralized management plane, a decoupled (from data) control plane, and a programmable data plane at customer premises. We present an ISP service delivery framework (iSDF) that provides ISPs a domain-specific API for network function virtualization by leveraging a programmable middlebox built from commodity home-routers. It also includes an application server to disseminate, configure, and update ISP services. We develop and report results for three diverse ISP applications that demonstrate the practicality and flexibility of iSDF, namely distributed VPN (control plane decisions), pay-per-site (rapid deployment), and BitTorrent blocking (data plane processing).

    Katerina Argyraki
  • A. Sivaraman, K. Winstein, P. Varley, J. Batalha, A. Goyal, S. Das, J. Ma, H. Balakrishnan

    In fields like data mining and natural language processing, design contests have been successfully used to advance the state of the art. Such contests offer an opportunity to bring the excitement and challenges of protocol design - one of the core intellectual elements of research and practice in networked systems - to a broader group of potential contributors, whose ideas may prove important. Moreover, it may lead to an increase in the number of students, especially undergraduates or those learning via online courses, interested in pursuing a career in the field. We describe the creation of the infrastructure and our experience with a protocol design contest conducted in MIT's graduate Computer Networks class. This contest involved the design and evaluation of a congestion-control protocol for paths traversing cellular wireless networks. One key to the success of a design contest is an unambiguous, measurable objective to compare protocols. In practice, protocol design is the art of trading off conflicting goals with each other, but in this contest, we specified that the goal was to maximize log(throughput/delay). This goal is a good match for applications such as video streaming or videoconferencing that care about high throughput and low interactive delays. Some students produced protocols whose performance was better than published protocols tackling similar goals. Furthermore, the convex hull of the set of all student protocols traced out a tradeoff curve in the throughput-delay space, providing useful insights into the entire space of possible protocols. We found that student protocols diverged in performance between the training and testing traces, indicating that some students had overtrained ("overfitted") their protocols to the training trace. Our conclusion is that, if designed properly, such contests could benefit networking research by making new proposals more easily reproducible and amenable to such "gamification," improve networked systems, and provide an avenue for outreach.

    Augustin Chaintreau
  • Darleen Fisher

    The Future Internet Architectures (FIA) constitutes a 10year effort by the U.S. National Science Foundation (NSF) that was launched in 2006, with the announcement of the Future INternet Design (FIND) research area within a Network Technologies and Systems (NeTS) program solicitation. This solicitation outlined a three-phase program to "rethink" the Internet, beginning with FIND and culminating in the recently announced two-year awards for Future Internet ArchitectureNext Phase (FIA-NP). Because many readers may not be familiar with the thinking behind this effort, this article aims to provide a historical context and background for the technical papers included in this issue.

    David Wetherall
  • D. Naylor, M. Mukerjee, P. Agyapong, R. Grandl, R. Kang

    Motivated by limitations in today's host-centric IP network, recent studies have proposed clean-slate network architectures centered around alternate first-class principals, such as content, services, or users. However, much like the hostcentric IP design, elevating one principal type above others hinders communication between other principals and inhibits the network's capability to evolve. This paper presents the eXpressive Internet Architecture (XIA), an architecture with native support for multiple principals and the ability to evolve its functionality to accommodate new, as yet unforeseen, principals over time. We present the results of our ongoing research motivated by and building on the XIA architecture, ranging from topics at the physical level ("how fast can XIA go") up through to the user level.

  • T. Wolf, J. Griffioen, K. Calvert, R. Dutta, G. Rouskas, I. Baldin, A. Nagurney

    The Internet has been a key enabling technology for many new distributed applications and services. However, the deployment of new protocols and services in the Internet infrastructure itself has been sluggish, especially where economic incentives for network providers are unclear. In our work, we seek to develop an "economy plane" for the Internet that enables network providers to offer new network-based services (QoS, storage, etc.) for sale to customers. The explicit connection between economic relationships and network services across various time scales enables users to select among service alternatives. The resulting competition among network service providers will lead to overall better technological solutions and more competitive prices. In this paper, we present the architectural aspects of our ChoiceNet economy plane as well as some of the technological problems that need to be addressed in a practical deployment.

  • A. Afanasyev, J. Burke, L. Zhang, claffy, L. Wang, V. Jacobson, P. Crowley, C. Papadopoulos, B. Zhang

    Named Data Networking (NDN) is one of five projects funded by the U.S. National Science Foundation under its Future Internet Architecture Program. NDN has its roots in an earlier project, Content-Centric Networking (CCN), which Van Jacobson first publicly presented in 2006.1 The NDN project investigates Jacobson's proposed evolution from today's host-centric network architecture (IP) to a data-centric network architecture (NDN). This conceptually simple shift has far-reaching implications for how we design, develop, deploy, and use networks and applications. We describe the motivation and vision of this new architecture, and its basic components and operations. We also provide a snapshot of its current design, development status, and research challenges. More information about the project, including prototype implementations, publications, and annual reports, is available on named-data.net.

  • A. Venkataramani, J. Kurose, D. Raychaudhuri, K. Nagaraja, M. Mao, S. Banerjee

    MobilityFirst is a future Internet architecture with mobility and trustworthiness as central design goals. Mobility means that all endpoints - devices, services, content, and networks - should be able to frequently change network attachment points in a seamless manner. Trustworthiness means that the network must be resilient to the presence of a small number of malicious endpoints or network routers. MobilityFirst enhances mobility by cleanly separating names or identifiers from addresses or network locations, and enhances security by representing both in an intrinsically verifiable manner, relying upon a massively scalable, distributed, global name service to bind names and addresses, and to facilitate services including device-to-service, multicast, anycast, and context-aware communication, content retrieval, and more. A key insight emerging from our experience is that a logically centralized global name service can significantly enhance mobility and security and transform network-layer functionality. Recognizing and validating this insight is the key contribution of the MobilityFirst architectural effort.

  • T. Anderson, K. Birman, R. Broberg, M. Caesar, D. Comer, C. Cotton, M. Freedman, A. Haeberlen, Z. Ives, A. Krishnamurthy, W. Lehr, B. Loo, D. Mazières, A. Nicolosi, J. Smith, I. Stoica, R. van Renesse, M. Walfish, H. Weatherspoon, C. Yoo

    NEBULA is a proposal for a Future Internet Architecture. It is based on the assumptions that: (1) cloud computing will comprise an increasing fraction of the application workload offered to an Internet, and (2) that access to cloud computing resources will demand new architectural features from a network. Features that we have identified include dependability, security, flexibility and extensibility, the entirety of which constitute resilience. NEBULA provides resilient networking services using ultrareliable routers, an extensible control plane and use of multiple paths upon which arbitrary policies may be enforced. We report on a prototype system, Zodiac, that incorporates these latter two features.

  • P. Bosshart, D. Daly, G. Gibb, M. Izzard, N. McKeown, J. Rexford, C. Schlesinger, D. Talayco, A. Vahdat, G. Varghese, D. Walker

    P4 is a high-level language for programming protocol-independent packet processors. P4 works in conjunction with SDN control protocols like OpenFlow. In its current form, OpenFlow explicitly specifies protocol headers on which it operates. This set has grown from 12 to 41 fields in a few years, increasing the complexity of the specification while still not providing the flexibility to add new headers. In this paper we propose P4 as a strawman proposal for how OpenFlow should evolve in the future. We have three goals: (1) Reconfigurability in the field: Programmers should be able to change the way switches process packets once they are deployed. (2) Protocol independence: Switches should not be tied to any specific network protocols. (3) Target independence: Programmers should be able to describe packet processing functionality independently of the specifics of the underlying hardware. As an example, we describe how to use P4 to configure a switch to add a new hierarchical label.

    Marco Mellia
  • G. Kalejaiye, J. Rondina, L. Albuquerque, T. Pereira, L. Campos, R. Melo, D. Mascarenhas, M. Carvalho

    This paper describes a strategy that was designed, implemented, and presented at the Mobile Ad Hoc Networking Interoperability and Cooperation (MANIAC) Challenge 2013. The theme of the challenge was "Mobile Data Offloading" and consisted on developing and comparatively evaluating strategies to offload infrastructure access points via customer ad hoc forwarding using handheld devices. According to the challenge rules, a hop-by-hop bidding contest should decide the path of each data packet towards its destination. Consequently, each team should rely on other teams' willingness to forward packets for them in order to get their traffic across the network. Following these rules, this paper proposes a strategy that is based on the concept of how "tight" a node is to successfully deliver a packet to its destination within a given deadline. This "tightness" idea relies on a shortest-path analysis of the underlying network graph, and it is used to define three sub-strategies that specify a) how to participate in an auction; b) how to announce an auction; and c) how to decide who wins the announced auction. The proposed strategy seeks to minimize network resource utilization and to promote cooperative behavior among participant nodes.

    Sanjay Jha
  • T. Krenc, O. Hohlfeld, A. Feldmann

    On March 17, 2013, an Internet census data set and an accompanying report were released by an anonymous author or group of authors. It created an immediate media buzz, mainly because of the unorthodox and unethical data collection methodology (i.e., exploiting default passwords to form the Carna botnet), but also because of the alleged unprecedented large scale of this census (even though legitimate census studies of similar and even larger sizes have been performed in the past). Given the unknown source of this released data set, little is known about it. For example, can it be ruled out that the data is faked? Or if it is indeed real, what is the quality of the released data? The purpose of this paper is to shed light on these and related questions and put the contributions of this anonymous Internet census study into perspective. Indeed, our findings suggest that the released data set is real and not faked, but that the measurements suffer from a number of methodological flaws and also lack adequate meta-data information. As a result, we have not been able to verify several claims that the anonymous author(s) made in the published report. In the process, we use this study as an educational example for illustrating how to deal with a large data set of unknown quality, hint at pitfalls in Internet-scale measurement studies, and discuss ethical considerations concerning third-party use of this released data set for publications.

  • C. Davids, G. Ormazabal, R. State

    In this article we describe the discussion and conclusions of the "Roundtable on Real-Time Communications Research: What to Study and How to Collaborate" held at the Illinois Institute of Technology's Real-Time Communications Conference and Expo, co-located with the IPTComm Conference, October 15-17, 2013.

  • kc claffy, D. Clark

    On December 12-13 2013, CAIDA and the Massachusetts Institute of Technology (MIT) hosted the (invitation-only) 4th interdisciplinary Workshop on Internet Economics (WIE) at the University of California's San Diego Supercomputer Center. This workshop series provides a forum for researchers, commercial Internet facilities and service providers, technologists, economists, theorists, policy makers, and other stakeholders to inform current and emerging regulatory and policy debates. The theme for this year's workshop was the economic health of the Internet ecosystem, including emphasis on the cost of and revenue sources to support content delivery, the quality of user experience, economic and policy influences on and effects of emerging specialized services, and the role of data in evaluating ecosystem health. This report describes the discussions and presents relevant open research questions identified by participants. Slides presented at the workshop and a copy of this final report are available at http://www.caida.org/workshops/wie/1312/.

Syndicate content