Dirk Kutscher

Personal web page

Managing Radio Networks in an Encrypted World

without comments

I attended last week’s IAB/GSMA Workshop on Managing Radio Networks in an Encrypted World (MaRNEW).

The motivation for this workshop was the increasing trend of applying transport layer end-to-end encryption in major web applications such as Google services, YouTube, Netflix, Facebook and others. This trend will likely increase due to further deployment of HTTP/2 for which client implementations today try to setup TLS connections per default.

In mobile networks, traffic management but also additional services/functions have traditionally relied on being able to leverage knowledge about application type, application specifics. Example for such functions include policing/prioritization, optimized scheduling, caching, filtering, but also tracking, ad-insertion etc. In addition to functions that operators want to apply, there are also regulation requirements (depending on local legislation) for filtering, legal intercepting etc. that would become more difficult in the presence of ubiquitous encryption.

At the MaRNEW workshop, leading experts from network operators, vendors, application service providers, CDN providers and academic institutions discussed the impact of ubiquitous encryption as well as ideas for enabling an effective collaboration between the network, applications and users to enable optimal performance and resource efficiency.

In particular, the workshop addressed the following topics:

  • Understanding the bandwidth optimization use cases particular to radio networks;
  • Understanding existing approaches and how these do not work with encrypted traffic;
  • Understanding reasons why the Internet has not standardised support for legal interception and why mobile networks have;
  • Determining how to match traffic types with bandwidth optimization methods;
  • Discussing minimal information to be shared to manage networks but ensure user security and privacy;
  • Developing new bandwidth optimization techniques and protocols within these new constraints;
  • Discussing the appropriate network layer(s) for each management function; and
  • Cooperative methods of bandwidth optimization and issues associated with these.

Encryption: Technological and Business Aspects

It is not a secret that there are different aspects for discussing end-to-end encryption in public networks. Obviously, encryption helps with user privacy, and with the background of recent and current revelations of privacy breaches through pervasive monitoring, it has become common agreement that more (easily deployable) encryption would be useful to overcome this.

There is however also the business perspective: the Internet and specifically the eco system of mobile communication and service provision has multiple stake holders, each of those with their particular interests: network operators want to provide a useful service, in an economical way and may have an interest to enhance the overall service quality through various technical measures. Application service providers want their particular service to perform well over a range of different networks. Network equipment vendors have their product roadmaps and network architecture preferences etc.

Finally, there are the actual users of the system who have an interest in good quality of experience, cost-efficiency — and privacy. Privacy is not only a concern with respect to (illegal) pervasive monitoring by agencies, but also with respect to maintaining anonymity and confidentiality towards network and service providers. For many applications, user profiles, user-generated data etc. is also a key business asset — so there is a strong interest by different players to either get access to that data — or (depending on the nature of a player) to keep other players from accessing it — through encryption.

The MaRNEW workshop focused on the technological discussion.

Impact of Encryption

During the discussion the following main impacts of ubiquitous encryption on mobile network were identified:

  • Traditional ways of identifying and classifying network traffic (DPI) become more costly and potentially infeasible.
  • Traditional traffic management systems have relied on such classification, for different purpose: optimizing resource usage in access networks according to operator policies, forwarding of traffic through optimizers, caches etc., as well as filtering. Those approaches and the actual requirements behind them need to be revisited.
  • Content and service provisioning in both mobile and fixed networks today is heavily relying on CDN and in-network application functions. In addition, new approaches such as Mobile Edge Computing may shift more of such functions to access networks. The motivation is to provide better performance and cost efficiency through offloading networks (CDN cache hits) and through reducing latency and transport protocol performance (local control loops, reduced RTT to caches). Introducing more and more end-to-end encryption makes it impossible for operators to provide any application (or CDN-provider)-independent optimization functions. The alternative of running individual instances for each individual CDN provider does not seem promising. It could also be a major road block for future network and application innovation — because each of those individual functions might require upgrading to introduce in-network support for it.

Way Forward

cooperative-traffic-management

 

(Copyright 2015 NEC)

At the workshop, different solutions were discussed.

  • First, it was agreed that the actual impact needs to be understood better and ought to be quantified. For example, assuming that some knowledge about application types (or corresponding service quality expectations) could be leveraged by base stations for more efficient transmission scheduling (e.g., by delaying packets of non-latency-sensitive flows or by operating multiple queues for different flow types), networks should at least be able to obtain corresponding hints from senders. However, the actual impact and potential benefits have to be demonstrated. Operators will work on that issue.
  • The (Internet) transport protocol community has made significant progress in recent years on several fronts: Active Queue Management (AQM) such as fq_codel and PIE have been demonstrated to be able to improve load balancing and reduce latency in router queues. Moreover, transport protocol research has led to promising results (for example PCC — Performance-oriented Congestion Control). It was suggested that those mechanisms should be implemented and deployed where possible.
  • Several options for Cooperative Traffic Management have been discussed. For example this could included exchanging certain information between the network and senders/receivers. The network could inform endpoints better about congestion and non-congestion-induced problems (for example in an extended ECN fashion), or endpoints could inform the network about relevant meta information (application type, QoS requirements etc.). The latter could leverage existing technologies such as DiffServ. Potentially, it would be sufficient to distinguish delay-sensitive flows (e.g., for interactive real-time) and delay-tolerant flows (file download etc.). One interesting question is how endpoints would be incentivized to use such signaling correctly and how corresponding APIs would look like.
  • Overcoming the general limitations of connection-based security and its tendency to require application-specific (or CDN-provider-specific) in-network functions could require a more fundamental rethinking of network architecture and protocol layering. For example, Information-Centric Networking (ICN) would leverage object-security (authentication, encryption), hence enabling the network to implement functions such as caching, local transport strategies etc. in an application manner. This could be of particular relevance for 5G networks where a higher level of dynamicity in the creation and deployment of new OTT services are expected.

For the discussion of such solutions, I (together with several colleagues) have made two contributions to the workshop: 1) Enabling Traffic Management without DPI, and 2) Maintaining Efficiency and Privacy in Mobile Networks through Information-Centric Networking.

Enabling Traffic Management without DPI

Is DPI really needed for traffic management in mobile networks? Our position is “no”. Traffic management is usually realized through relatively simple mechanisms like rate shaping, prioritization, and dropping packets. Compared to these mechanisms, the semantics of applications that can be exposed through DPI are much richer; traffic classification anyway maps these semantics down to a simple set of categories.

The question then arises whether operators are really helped by brittle, insecure and expensive mechanisms for gaining higher fidelity information for the coarse traffic information for traffic management, or whether simple signaling would suffice for traffic classification for mobile network management purposes.

Obviously, when relying on endpoints to signal information about the underlying application which may be used to change the network’s treatment of that application’s traffic, questions of trust arise: how can the network be sure the endpoints are being honest, and prevent endpoints from gaming the system to their advantage (and the disadvantage of others); can these signaling approaches be used as an attack vector. Here the approach is to define the vocabulary of the signaling protocol to properly incentivize honest cooperation, while allowing the network to verify this cooperation.

We discuss two application-independent approaches for traffic management that are based on network-compatible metrics: ConEx Policing and low latency support with SPUD.


Congestion Exposure (ConEx) is a mechanism that enables senders to inform the network about previously encountered congestion in flows thus enabling senders and network infrastructure to respond to congestion based on operator policies. This information is provided in the IP header and can still be accessed even if the payload is encrypted. ConEx information is auditable by comparing the congestion level at network egress to the ConEx signal which incentivizes the sender to state its congestion contribution correctly.

Using ConEx would allow for a bulk packet traffic management system that does not have to consider application classes. Instead, with ConEx accurate downstream path information on incipient congestion are visible to ingress network operators. This information can be used to base traffic management on the actual current cost (which is the contribution to congestion of each flow) and enable operators to apply congestion-based policing/accounting depending on their preference and independent of application characteristics. Such traffic management would be simpler, more robust (no real-time flow application type identification required, no static configuration of application classes) and provide better performance as decisions can be taken based on the real actual cost contribution at each point in time.

The Substrate Protocol for User Datagrams (SPUD) is a new approach to selective information exposure designed to support transport evolution. SPUD is realized as a shim between UDP and an (encrypted) transport protocol. The basic SPUD protocol provides minimal sub-transport functionality by grouping of packets together into tubes and signaling of the start and end of a tube.

This will assist middleboxes in state setup and teardown along the path. Further, SPUD provides an extensible signaling mechanism based on a type-value encoding for associating properties with individual packets or all packets in a tube. The SPUD protocol can be used to signal low latency requirements from an endpoint to the network, or expose the existence of support for such services from the network to the endpoint. Therefore we propose to provide four SPUD signals: a latency sensitivity flag, a signal to yield to another tube, an application preference for a maximum single queue delay, and a facility to discover the maximum possible single queue length along the path.

Based on the latency-sensitivity flag a network operator can implement an additional service (as compared to today’s best effort service) that uses smaller queues and/or different AQM parameters without changing the service that is provided today. Signaling of lower queue priority or maximum single hop delay can further be used to preferentially drop packets of the same sender or within one flow. Information about expected queuing delays on the path can be used for buffer configuration at the endpoints.

The proposal is not intended as a blueprint for immediate implementation — but it demonstrates how cooperative traffic management could be implemented. In our view, cooperative traffic management requires a solid understanding of the interactions with transport layer and the corresponding performance impacts/improvements.

Maintaining Efficiency and Privacy in Mobile Networks through Information-Centric Networking

We present a solution to overcome the impasse of deploying confidentiality at the cost of breaking most of current network traffic engineering in mobile networks. Our proposition is based on Information-Centric Networking (ICN) which is a data-centric network architecture that gracefully incorporates security and traffic optimization.

Content-based security instead of connection based is the foundation of the Information-Centric Networking (ICN) architecture. In ICN, we provide a network service that directly implements the desired information-access abstraction. The network forwards requests for named data and corresponding responses containing the data. The name can be cryptographically bound to the data for ascertaining authenticity. This enables the network to replicate data objects in arbitrary locations, thus enabling ubiquitous caching. Object data can also be encrypted for user privacy, leaving other network-relevant information such as the name intact – thus maintaining options for traffic management, policing etc. The performance gains of having ICN in the mobile backhaul have been evaluated experimentally (see paper). ICN incorporates these ideas into a novel network layer providing all of the mentioned objectives without using man-in-the-middle like solutions.

ICN secures data itself by requiring producers to cryptographically sign every data packet: the signature constitutes the integrity meta-data. The data is uniquely identified by a name that is bound to the data via the signature. The producer’s public key to implement signature verification can be obtained by using the KeyLocator field which can be the name of the data containing the key of the producer. Authentication is implemented via the producer’s key that makes use of a trust model, e.g. PKI, Web-of-Trust that can be extended using key chaining to delegate trust to different sub-namespaces (for hierarchical naming). Confidentiality is obtained by encryption of the data payload using the producer’s key. Notice that authenticity, integrity and confidentiality are independent features.

Once data is published by the producer it can be stored in any location without affecting the security properties of the data which are location independent. Inter-networking of encrypted data is included by design in ICN and in-network caching is always possible with or without confidentiality. Authenticity might not be necessary in many cases so the authentication of the identity of the producer is optional. It is not mandatory either to verify the integrity of the data by verification of the signature. It is important to remark that ICN disantangles authenticity, privacy and integrity so that they can be handled in different ways and without the interaction of end-hosts.

TLS provides web security by encrypting a layer 4 connection between two hosts. Authenticity is provided by the web of trust (certification authorities and a public key infrastructure) to authenticate the web server and symmetric cypher on the two end points based on a negotiated key. In presence of TLS many networking operations become unfeasible: filtering, caching, acceleration, trans-coding.

ICN takes a radically different approach to guarantee confidentiality, authenticity and integrity by embedding them into a redefined network layer. Indeed, ICN builds on the abstraction of data requested, accessed, cached and forwarded by name: the network forwards requests coming from the consumer for named data and routes back data packets on the identical reverse path (symmetric routing).

The ICN communication model allows network nodes between a web server and a web client to operate as forwarding and storage functions to implement various inter-networking functionalities like caching or load balancing without relaxing any security feature. As a fully fledged data-centric network architecture, ICN incorporates mobility, storage, security and multi-point communication by design.

Written by dkutscher

September 28th, 2015 at 12:49 am

ICN-2015 Conference Program

without comments

Join us for the ICN-2015 Conference in San Francisco from Sep. 30 to Oct. 2.

ACM ICN is an annual conference of the ACM Special Interest Group on Data Communication (SIGCOMM) on information-centric networking.

In a nutshell, this year’s conference includes
– 1 keynote given by Van Jacobson
– 19 full papers presented in single track format
– 8 posters
– 10 demos
– 2 full-day tutorials
– 1 industrial panel

Conference details:
http://conferences.sigcomm.org/acm-icn/2015/

Registration details:
http://www.regonline.com/icn2015

Keynote:
– Van Jacobson, Internet pioneer and core architect of Named Data
Networking (NDN), will talk about “Improving the Internet with ICN”.

Tutorials:
– CCN: Practical CCNx – Protocol and Code
– NDN: Security & Synchronization in Named Data Networking (NDN)

Panel:
– Next Steps for ICN: Research, Applications, Deployment and Economics

Topics of papers, posters, and demos include:
– Architecture design and evaluation
– Comparison of ICN architecture proposals
– Limits and limitations of ICN architectures
– ICN evaluation methodology and metrics
– Evaluation of ICN benefits
– Analysis of scalability issues in ICN
– ICN enabled applications
– Routing in ICN networks
– Mobility support
– Trust management
– Access control mechanisms
– ICN economics and business models
– Tools and experimentation facilities
– Measurement methodologies
– Experience from implementations and experiments
– Specific scenarios and implementation approaches
– Feasibility studies for high speed networking
– Privacy
– ICN Deployment
– ICN APIs

Check out the program.

Written by dkutscher

August 20th, 2015 at 10:42 am

Posted in Events

Tagged with ,

Privacy, Performance, Protocols: ICN Researchers meet in Prague

without comments

Prague

The IRTF Information-Centric Networking Research Group (ICNRG) had another ICN research fest with two meetings this week in Prague where IETF-93 is taking place.

ICN is an approach to evolve the Internet infrastructure to directly support information distribution by introducing uniquely named data as a core Internet principle. Data becomes independent from location, application, storage, and means of transportation, enabling in-network caching and replication. This enables the design of more robust, secure and better performing networked systems.

This week, more than 100 researchers got together to discuss recent advances in protocol development, performance optimizations, user privacy and new use cases.

ICNRG meeting

One of the protocols that are developed in ICNRG is CCNx, a network protocol that provides requests (Interests) for named data and Content Object responses. The protocol semantics are specified in draft-irtf-icnrg-ccnxsemantics, and the protocol format is specified in draft-irtf-icnrg-ccnxmessages. The ICN community is currently discussing several extensions to the protocol, including support for “manifest” objects, which would facilitate the distribution of larger, chunked objects and add additional performance and flexibility to ICN systems.

Another highlight of the meeting was a presentation by Iannis Psaras from UCL on Solving the Congestion Problem using ICNPrinciples, an approach that is using Resource Pooling as a tool to manage uncertainty in congestion management.

Vasilis Sourlas (UCL) presented Information Resilience through User-Assisted Caching in Disruptive Content-Centric Networks. The corresponding paper won the IFIP 2015 best paper award and describes work from the GreenICN project. The approach relies on a modified NDN router design that features a “Satisfied Interest Table” (SIT) that enables user-assisted caching.

Bengt Ahlgren (SICS) presented on the Applicability and Tradeoffs of ICN for Efficient IoT (draft-lindgren-icnrg-efficientiot). This document outlines the tradeoffs involved in utilizing Information Centric Networking (ICN) for the Internet of Things (IoT) scenarios. It describes the contexts and applications where the IoT would benefit from ICN, and where a host-centric approach would be better. The requirements imposed by the heterogeneous nature of IoT networks are discussed (e.g., in terms of connectivity, power availability, computational and storage capacity). Design choices are then proposed for an IoT architecture to handle these requirements, while providing efficiency and scalability. An objective is to not require any IoT specific changes of the ICN architecture per se, but we do indicate some potential modifications of ICN that would improve efficiency and scalability for IoT and other applications.

Dirk Trossen (Inter Digital) presented IPoverICN – the Better IP?, a presentation of the EU-H2020 POINT project that is developing an IP over ICN system. The hypothesis of this project is that IPoverICN has the potential to run IP services better than in standard IP networks.

Mark Stapp (Cisco) presented on Private Communication in ICN. This presentation asks the question whether ICN needs better privacy protection to achieve parity with IP for user privacy in the presence of ubiquitous encryption. The discussion initiated an intensive discussion on privacy requirements for ICN that will continue in upcoming meetings.

Jan Seedorf (NEC) presented on Using ICN in Disaster Scenarios (draft-seedorf-icn-disaster/). This is a presentation of the GreenICN project and summarized some research challenges for coping with natural or human-generated, large-scale disasters. Further, the document discusses potential directions for applying Information Centric Networking (ICN) to address these challenges.

All presentations and detailed notes can be found at the ICNRG Wiki.

This summer will host a series of additional ICN events:

  • ACM SIGCOMM ICN 2015 Conference in San Francisco (September 30 — October 2)
  • NDN Community meeting at UCLA (September 28 — 29)
  • ICNRG Interim Meeting in Palo Alto (October 3)
  • Written by dkutscher

    July 23rd, 2015 at 1:13 pm

    Posted in Posts

    Tagged with , , ,

    Open Source Carrier Networking

    without comments

    Open Source Software development models are changing the way the telco industry is creating products and systems. This presentation at ONS-2015 discusses how innovation, agile development and Open Source Software are linked together.It presents experience with transforming telco vendor development from closed to open source and provides an outlook of future activities in the NFV space.

    Talk Info (Presentation available on request)

    Written by dkutscher

    June 18th, 2015 at 11:22 pm

    Posted in Talks

    Tagged with , , , , ,

    OPNFV Arno Released

    without comments

    The OPNFV project has released its first major software release “Arno”.

    OPNFV is a carrier-grade,integrated,  open source platform to accelerate the introduction of new Network Functions Virtualization (NFV) products and services.

    Arno is a developer-focused release that provides an initial build of the NFV Infrastructure (NFVI) and Virtual Infrastructure Manager (VIM) components of ETSI NFV architecture.

    Key capabilities of OPNFV Arno:

    • Availability of baseline platform: Arno enables continuous integration, automated deployment and testing of components from upstream projects such as Ceph, KVM, OpenDaylight, OpenStack and Open vSwitch. It allows developers and users to automatically install and explore the platform.
    • Ability to deploy and test various VNFs: End users and developers can deploy their own or third party VNFs on Arno to test its functionality and performance in various traffic scenarios and use cases.
    • Availability of test infrastructure in community-hosted labs: Agile testing plays a crucial role in the OPNFV platform. With Arno, the project is unveiling a community test labs infrastructure where users can test the platform in different environments and on different hardware. This test labs infrastructure enables the platform to be exercised in different NFV scenarios to ensure that the various open source components come together to meet vendor and end user needs.
    • Allows automatic continuous integration of specific components: As upstream projects are developed independently they require testing of various OPNFV use cases to ensure seamless integration and interworking within the platform. OPNFV’s automated toolchain allows continuous automatic builds and verification.

     

    Links

     

    Written by dkutscher

    June 4th, 2015 at 6:12 pm

    Posted in Projects

    Tagged with , , , ,

    The Next Step of OpenStack Evolution for NFV Deployments

    without comments

    Chris Wright and I presented on “The Next Step of OpenStack Evolution for NFV Deployments” at last week’s OpenStack Summit in Vancouver.


    Presentation at OpenStack Summit

    NFV is now a well-known concept and in an early deployment stage, leveraging and adapting OpenStack and other Open Source Software systems. In the OPNFV project, a large group of industry peers is building a carrier-grade, integrated, open source reference platform for the NFV community. The telco industry has successfully adopted Open Source Software for carrier-grade deployments. It is now time for taking the next steps and to extend the colloaboration with upstream projects — by opening up previously proprietary developments, by contributing code and other artifacts in order to create a ecosystem of NFV platforms, applications, and management/orchestration systems.

    This presentation shares some insights on how Red Hat and NEC are working together to foster collaboration in the NFV ecosystem by actively working with OpenStack and other upstream projects.

    NEC has pioneered the adoption of Linux, KVM, Open vSwitch, and OpenStack for their mobile network core product line (virtualized EPC)
    and has gained significant experience through development work and deployments. NEC’s extensions for high efficiency and high
    availability have led to contributions of new features to OpenStack, such as DPDK vSwitch control and CPU allocation features. For NEC, it is very important to have those features integrated into the mainstream code base for building reliable infrastructure systems.

    Red Hat, one of main contributors to OpenStack, leads the development of those functions to meet NFV requirements in OpenStack, making critical and demanding applications run of top of open platforms. The presentation explains how NEC and Red Hat are integrating and optimizing Red Hat Enterprise Linux OpenStack Platform and NFV, along with contributions to open source communities, including OpenStack and Open Platform for NFV (OPNFV).

    Written by dkutscher

    May 26th, 2015 at 11:25 pm

    Posted in Talks

    Tagged with , , ,

    Scalable Content Exchange in Challenged ICNs

    without comments

    I presented GreenICN work on Scalable Content Exchange in Challenged ICNs at CCNxCon-2015 this week.

    Download: ccnxcon2015-kutscher.pdf

    Abstract:
    The principles of Information­Centric Networking (ICN), accessing data objects by name (not by location address), securing data objects (not connections), in­network caching (for sharing, repair, rate adaptation) make ICN attractive for a wide range of application scenarios beyond traditional data center or telco access network scenarios. In fact, one of the first instantiation of ICN had been developed based on Delay­Tolerant Networking (DTN) technologies.

    Currently, ICN/DTN is considered a promising approach for enabling/enhancing communication in disaster scenarios. In such scenarios, so­called ICN data mules (that carry and disseminate data times) may move randomly, and each time data mules encounter one another exchange data items. We envision that in such a scenario where there is no connectivity, data mules (e.g. vehicles or drones) can move around randomly. So these mobile routers interact with end users, working base stations and other data mules to fetch and deliver the data and queries. Thus, we do not consider adhoc networks where you can build a path to the destination reactively or proactively, rather a DTN like scenario.

    Consider a large scale disaster scenario like the earthquake in Japan in 2011 , where people in different parts of the city are stranded without the internet connectivity. But there are some zones, where base stations are still working and providing connectivity. Essentially, the scenario is such that ICN data mule move randomly across a geographic area, and when meeting end­users receive interests from them and also forward corresponding data items to end­users (if present in the content store / cache of the data mule). At the same time, when data mules encounter each other, they forward to each other certain end­user interest and/or data items (according to a predefined rule set and algorithm), such that interests and data items can be forwarded in a hop­by­hop DTN fashion. One research problem in such a scenario is how to optimize such data exchanges among data mules for optimal data dissemination (e.g. optimizing how many desired messages reach their recipients within a given timeframe with a given forwarding strategy, assuming that data mules only have limited time at each encounter to exchange
    messages).

    Written by dkutscher

    May 21st, 2015 at 4:57 pm

    Posted in Talks

    Tagged with , , , ,

    ICN Researchers Meet in Cambridge, MA

    without comments


     

    The ICN Research Group of the IRTF has met for a two day meeting in Cambridge, MA on January 13/14. More than 30 researchers from the US, Europe, China, and Japan gathered to discuss hot research topics in ICN such as:

    • Native ICN-based video streaming
    • Security (authenticated denial in ICN)
    • IoT and ICN
    • Hop-by-hop control messages in CCN
    • Named Function Networking

    In addition, different groups presented updates on their current implementations and their design decisions for packet formats and ICN protocols. For CCN-based protocols further steps towards a common format have been made.

    The next meeting (planned for the week of March 23rd in Dallas, at IETF-92) will continue the packet format discussion and progress new topics such as Named Function Networking.

     

    Written by dkutscher

    January 15th, 2015 at 3:42 pm

    Posted in Events

    Tagged with , ,

    Call for Papers: 2nd ACM Conference on Information-Centric Networking (ICN 2015)

    without comments

    ACM ICN 2015, September 30 - October 2, 2015, San Francisco, USA

    The Call for Papers for the 2nd ACM Conference on ICN is out:

     

                             Call for Papers

    ** 2nd ACM Conference on Information-Centric Networking (ICN 2015) **

     

    Sponsored by ACM and ACM SIGCOMM

     

    http://conferences.sigcomm.org/acm-icn/2015

     

    San Francisco, USA, September 30 – October 2, 2015

     

     

    Information Centric Networking (ICN) is a new network architecture intended to provide access to information without requiring an explicit binding of that information to a particular location. By directly addressing information, ICN supports mobile users and mobile networked devices, offers a higher-level communication service to applications, and promotes authentication and efficiency in the transmission and dissemination of information. Over the last few years, a global research and development community has grown around the idea of ICN.

     

    ACM ICN 2015 is the second edition of the ACM Conference on Information-Centric Networking, which follows a series of workshops on ICN held in conjunction with the ACM Sigcomm conference.  ACM ICN 2015 is the premier international forum for researchers and practitioners to present and discuss the most recent innovations, trends, experiences, and challenges in information centric networking.  ACM ICN 2015 will be a single-track conference featuring paper and poster presentations, panel discussions, and demonstrations.

     

    The Technical Program Committee of ACM ICN 2015 invites high-quality submissions describing unpublished research results in all aspects of ICN, with particular emphasis on contributions to architectural designs and reproducible experimental evaluations.  Papers submitted for consideration should not have been already published elsewhere and should not be under review or submitted for review elsewhere during the consideration period.

    Specifically, authors are required to adhere to the ACM Policy and Procedures on Plagiarism

    (http://www.acm.org/publications/policies/plagiarism_policy) and the ACM Policy on Prior Publication and Simultaneous Submissions (http://www.acm.org/publications/policies/sim_submissions).

     

    Topics of interest include:

     

    * Architecture design and evaluation

    * Comparison of different ICN architectures

    * Interoperability across ICN architectures

    * ICN evaluation methodology and metrics

    * Analysis of scalability issues in ICN

    * ICN enabled applications

    * Routing in ICN

    * Transport issues in ICN

    * Caching

    * Mobility support

    * Trust management and access control

    * Management in ICN

    * ICN economics and business models

    * Tools, experimentation facilities, and measurement methodology for ICN

    * Experience from implementation

    * Feasibility studies of ICN for high speed networking

    * Privacy

    * ICN Deployment

    * ICN APIs

     

    ———————–

    Submission Instructions

    ———————–

     

    Submitted papers can be up to 10 pages in length following the SIGCOMM format. All submissions must be in English and in PDF format. Submissions that do not comply with these instructions will be rejected without review.

    Papers must be submitted electronically through the ICN 2015 submission site.

     

    Submissions will be reviewed and evaluated on the basis of originality, importance of contribution, soundness, evaluation, quality of presentation and appropriate comparison to related work. The program committee as a whole will make final decisions about which submissions to accept for presentation at the conference. The program committee may propose that authors present their work with a poster accompanied by a 2-page extended abstract. ACM ICN 2015 also invites proposals for demos, tutorials and panel sessions.

     

    —————

    Important Dates

    —————

     

    Full Paper Submission: May 22, 2015

    Acceptance Notification: July 20, 2015

    Camera Ready Due: Aug. 15, 2015

    Conference: September 30 – October 2, 2015

     

    ————————-

    Conference General Chairs

    ————————-

     

    – Nacho (Ignacio) Solis (PARC, USA)

     

    ———————————-

    Technical Program Committee Chairs

    ———————————-

     

    – Antonio Carzaniga (USI, Switzerland)

    – K. K. Ramakrishnan (UC Riverside, USA)

     

    ———————————–

    Technical Program Committee Members

    ———————————–

     

    – Mayutan Arumaithurai (University of Goettingen, Germany)

    – Giuseppe Bianchi (University of Rome “Tor Vergata”, Italy)

    – Nicola Blefari-Melazzi (University of Rome “Tor Vergata”, Italy)

    – Jeff Burke (UCLA, USA)

    – Kenneth Calvert (University of Kentuky, USA)

    – Giovanna Carofiglio (Cisco)

    – Patrick Crowley (Washington University, USA)

    – Christian Esteve Rothenberg (UNICAMP, Brazil)

    – JJ Garcia-Luna-Aceves (University of California Santa Cruz, USA)

    – Toru Hasegawa (Osaka University, Japan)

    – Jussi Kangasharju (University of Helsinki, Finland)

    – Satyajayant Misra (New Mexico State University, USA)

    – Vishal Misra (Columbia University, USA)

    – Luca Muscariello (Orange Labs, France)

    – Kiran Nagaraja (Ericsson)

    – Dave Oran (Cisco, USA)

    – Jörg Ott (Aalto University, Finland)

    – Christos Papadopoulos (Colorado State University, USA)

    – Craig Partridge (BBN, USA)

    – Diego Perino (Alcatel Lucent, France)

    – George Polyzos (AUEB, Greece)

    – Yiannis Psaras (UCL, UK)

    – Dipankar Raychaudhuri (Rutgers University, USA)

    – Jim Roberts (IRT SystemX, France)

    – Dario Rossi (Telecom ParisTech, France)

    – Thomas Schmidt (HAW Hamburg, Germany)

    – Jan Seedorf (NEC Labs Europe)

    – Nacho (Ignacio) Solis (PARC, USA)

    – Karen Sollins (MIT, USA)

    – Christian Tschudin (Uni Basel, Switzerland)

    – Arun Venkataramani (UMass, USA)

    – Matthias Wählisch (FU Berlin, Germany)

    – Roy Yates (Rutgers University, USA)

    – Lixia Zhang (UCLA, USA)

     

     

    ————

    More Details

    ————

     

    Please see http://conferences.sigcomm.org/acm-icn/2015

     

    Written by dkutscher

    January 9th, 2015 at 9:57 am

    Posted in Events

    Tagged with , , ,

    Open Standards, Open Source, Open Loop

    without comments

    Dave Ward (Chief Architect at Cisco) gave an interesting lunch talk on the relationship between Open Standards and Open Source Software at IETF-91 today. Technologies such as OpenFlow and NFV are increasingly being advanced through Open Source Software projects that develop both individual components as well as larger systems. The developed artifacts are  sometimes being referred to as de-facto standards.

    Dave gave some perspectives on how Open Source Software can help to speed up collaborative technology development and related this to standards work in the IETF and other bodies. Dave emphasized the importance of Open Standards for the development of Internet technologies but he pointed out that Open Standards can leverage Open Source to speed up specification development and to validate architecture and protocol specifications.

    The talk suggested embracing Open Source Software development for standards work in the IETF, pointing at new working models and skill sets that were required for that.

    Obviously, the IETF has always had a focus on running code for validating specification and several recent efforts have been leveraging OSS succesfully — for example CORE, DTN and many others. Still, there was agreement that there is room for extending and potentially institutionalizing this.

     

     

     

     

    Written by dkutscher

    November 14th, 2014 at 4:37 am

    Posted in Events

    Tagged with , ,