Difference between revisions of "Real World Application Integration"

From Nsnam
Jump to: navigation, search
(some notes)
(Detail Schedule)
Line 48: Line 48:
== Detail Schedule ==
== Detail Schedule ==
* ~-4.30 read quagga source code, compile and run simple test; list simu_xxx sys-calls; finish the first step
* ~-4.30 read and compile quagga source code, list simu_xxx sys-calls;  
* 5.1~5.4 four days vacation
* 5.1~5.4 four days vacation
* ......
* 5.5~5.9 read zebra codebase, get some view of the code structure and running mechanism.
* 5.12....  work on a "netlink" socket for ns3.
== About Quagga ==
== About Quagga ==

Revision as of 14:52, 12 May 2008

Main Page - Current Development - Developer FAQ - Tools - Related Projects - Project Ideas - Summer Projects

Installation - Troubleshooting - User FAQ - HOWTOs - Samples - Models - Education - Contributed Code - Papers

Project Background

The goal of this proposed project is to develop frameworks and strategies that enable people to integrate already existing code into the simulator. Depending on the application, this can be a straightforward or laborious process, but still should be preferable in many cases to rewriting these protocols from scratch, and hopefully this project can come up with techniques to make these ports even easier.

Here are some initial pointers to how this problem has been worked in the past:

Gsoc Project

Liu Jian(liujatp@gmail.com) started the gsoc project from April 2008.


The purpose of the project is to develop frameworks and strategies that enable people to integrate already existing code into the ns-3 simulator, it will be accomplished by integrating Quagga, a routing deamon which implement many useful routing protocols. The project will begin by learning experence that Quagga ever ported to INET simulator, then porting Quagga to ns3 by adding some patchings, through the porting job, summariced,documented and structured properly, a adaption layer or some methodologies for ns-3 will be implemented, through which other real world application will be easily ported by the next person.

Project Plan

  • looking at quagga to identify the system calls it use.functions like socket, time, signal,etc.
  • implememt these function as simu_* in the ns-3-simu tree.
  • porting quagga to ns3.


  • listed all system functions which quagga calls. there is about 30 funs of total 150 need to be implemented in ns-3-simu.

After investigation, there were 4 types of functions.

  • next: implement these simu_xxx functions(there would be a great job)

started by simple server/client demo code from http://cs.ecs.baylor.edu/~donahoo/practical/CSockets/textcode.html running in ns-3-simu and test simu_x APIs.

  • .......

Detail Schedule

  • ~-4.30 read and compile quagga source code, list simu_xxx sys-calls;
  • 5.1~5.4 four days vacation
  • 5.5~5.9 read zebra codebase, get some view of the code structure and running mechanism.
  • 5.12.... work on a "netlink" socket for ns3.

About Quagga

basic knowledge

Quagga is a routing soute of 5 routing protocols(RIP,RIPng,OSPFv2,OSPFv3,BGP) based on Zebra, they can be run simutaneously or separately.Zebra layer that contains what is known as the "Routing Information Base" or RIB. Zebra is responsible for maintaining the RIB and for writing routes from the RIB into the kernel forwarding table.

Quagga was planned to use multi-threaded mechanism when it runs with a kernel that supports multi-threads. There may be several protocol-specific routing daemons and zebra the kernel routing manager.The ripd daemon handles the RIP protocol, while ospfd is a daemon which supports OSPF version 2. bgpd supports the BGP-4 protocol. For changing the kernel routing table and for redistribution of routes between different routing protocols, there is a kernel routing table manager zebra daemon. Quagga system architecture, see here http://www.quagga.net/docs/docs-info.php#SEC9.

see more information here:http://www.quagga.net/docs/docs-info.php

some notes

1, Firstly, if we donot care about how one of the protocols runs, we can get some idea about the main code-structure here, which would be useful for porting to ns3.

  • main structure. _ _ _the main thread maintains thead_master, which contains all 'threads' triggered from event, timer, I/O, background,etc; when all initial work was finished, the whole application run as event-driven mode like below:
  • thread. _ _ _Actually there were only one real-thread running at this time, it use the posix_signal hanlder functions, event functions of application-defined, many timer functions of the basic protocol-application , and select(2)system all to monitor I/O operations for multiplexing the events. the 'theads' mentioned above was not the common threads we commonly talk about, here it represent the all kinds of event-function defined above, the event-function, which was passed as function pointer and stored in thread-master, was called by sequence in main thread function depends on the different priority. the threads here has the _synchronous_ specific which was different from the threads in one real world application.
  • multithread mechanism in quagga. _ _ _As a real-world application, quagga deal with posix-signal from system kernel, I/O events from IO devices, application-level events from kernel code and timer functions of maitaining the application noraml-running.it seems that these kinds of events were asynchronous to each other, which looks like a multithreads-application. But here in quagga, a creative mechanism was used to avoid multithreads. A threadmaster was maintained in main-thread, which store all kinds of events mentioned above in the form of sigmaster(discuss below), timer-list, io(read,write)list, event-list,etc.In the main loop, only one thread with the highest priority was called in each loop, which has some similarity with ns3 simulator's event-schedule model. (the priority: posix_signal events > application-level events > timer events > IO events > background-timer events). so the whole application runs as only one real thread avoiding multithreads-application resource sharing problem and system asynchronously controling problem.
  • signal mechanism in quagga._ _ _As we know, quagga application can receive all signals from system and I/O, but it only defined some useful signal and trapped the others avoiding large system signal response. Usually, posix signal can be triggered at anytime from kernel, and it use the kernel execution stack and context, which was asynchronous to the main thread. To avoid these asynchronism, here, in quagga, A sigmaster was used, which stored the signal information instead of calling handler function immediately when one posix signal was triggered in kernel. Then, the main_thread check out all triggered signals of last cycle from the segmaster and call their handler functions. So, the posix_signals were processed _synchronously_ in the whole application.

2, started with Zebra.