|
Seminars /
2010-2011 June 08, 2011, 11:00 This seminar will be given in on Wednesday in collaboration with MLG The framework of this seminar is the problem of density estimation on
structured data (finite strings or trees). The most widely used model for
this kind of task is the Hidden Markov Model, or HMM, whose training is
done by the Baum-Welch algorithm, which maximizes the likelihood of a
training sample. It is then necessary to limit the expressiveness of the
considered HMM family by setting a maximum number of hidden states, in
order to avoid overfitting. This is usually done by an a priori knowledge,
or by some heuristics. Mai 27, 2011, 11:00 The Belgian Federal Police maintains a database with all recorded criminal facts in the country. For each fact that has been solved, one or several authors are known. One can then construct a network of crime authors by connecting authors to the crimes in which they are involved. In this seminar, I will present this unusual database and its particularities, and I will address the problem of community detection in this network. April 29, 2011, 11:00 Since the beginning of the nineties, the Internet has undergone impressive growth. This growth can be appreciated in terms of the equipment, such as routers and links, that has been added, as well as in the numbers of users and the value of commerce that it supports. In parallel to this expansion, over the past decade the networking research community has shown a growing interest in discovering and analyzing the internet topology. Some researchers have developed tools for gathering network topology data while others have tried to understand and model the Internet's properties. These efforts have brought us to a crucial juncture for toplogy measurement infrastructures: while, previously, these were both small (in terms of number of measurement points) and monolithic, we are starting to see the deployment of large-scale distributed systems composed of hundreds or thousands of monitors. As we look forward to this next generation of systems, we take stock of what has been achieved so far. In this survey, we discuss past and current mechanisms for discovering the internet topology at various levels: the IP interface, the router, and the AS level. In addition to discovery techniques, we provide insights into some of the well-known properties of the Internet topology. April 18, 2011, 10:00This seminar will be given on Monday at 10am Understanding the dynamics in large scale networks is a major challenge in front of the network research community. Traditional graph theoretic approaches have their own limitations and are not applicable due to the large size and dynamic nature of the network. In this background, my talk primarily addresses two different issues in technological and social networks. The first half of my talk is directed towards understanding the resilience and emergence of technological networks, specifically the superpeer networks. We propose an analytical framework in order to measure the resilience of superpeer networks in face peer churn and attacks. Other side, it is not obvious why bootstrapping of peer nodes and other local dynamics results in the appearance of bimodal degree distribution in superpeer networks like Gnutella. We develop a formalism which enables us to explain the emergence of bimodal network in face of dynamics like peer bootstrapping, churn, link rewiring etc. Further analysis leads us in formulating interesting bootstrapping protocols such that superpeer network evolves with desired topological properties.
April 07, 2011, 11:00 The problem of the shortest-path constitutes a central problem in graph theory and provides a formalism which can be used to describe and analyse numerous problems. For decades, the operations research community has been studying the intimate link that relates this theory to universal algebra. This connection is achieved through the use of algebraic structures such as semirings and dioids. The goal of this presentation is to underline these links and to provide pointers to various resources and possibilities provided by this particular perception of Euler's legacy. April 01, 2011, 11:00 When does a given finite set of linear operators possess a common invariant cone ? This question found important applications in recent papers on neural and gene networks, linear switched systems, functional equations, etc. In fact, linear operators with an invariant cone inherit many properties of nonnegative matrices (Perron-Frebenius theory, etc). For one operator this problem was solved in classical works of Krein, Rutman and Vandergraft. For several operators this is more difficult. We present a sharp criterion of the existence of a common invariant cone for an arbitrary set of matrices. It allows us, at least theoretically, to decide the existence of a common invariant cone just by comparing two special numbers depending on matrices. However, the computing of those numbers is hard. In fact, we show that this problem for arbitrary matrices with integer entries is algorithmically undecidable. On the other hand, some corollaries of the criterion lead to simple sufficient and necessary conditions. Finally, we formulate an approximative analogue of the problem and introduce a ``co-directional number'' of several matrices. Possible approaches to its computation are discussed. March 25, 2011, 11:00 Presentation of the results from her master thesis. March 18, 2011, 11:00 For undirected graphs, it is well known how to obtain the expected inter-vertex commute times from the graph Laplacian matrix. We show the same formulas hold in the case of strongly connected directed graphs. Our result is obtained by deriving a close relation between the Laplacian with a particular scaling and the so-called Fundamental matrix for an associated random walk over the graph. We find that the commute times still form a metric and give bounds in terms of the stationary probabilities for the random walk. We compare these commute times with those obtained from a previously proposed symmetrized Laplacian derived from a related weighted undirected graph. March 07, 2011, 12:00 This seminar will be given on Monday Our contemporary societies rely more and more on a steady and reliable power supply for their well-functioning. During the last few decades a number of large-scale power blackouts have been witnessed around the world, and this has caused major concerns among politicians and citizens. In this talk we will mention a few major power blackouts and discuss the sequence of events and why they occurred. These empirical examples show that major power blackouts often are results of a cascading of failures (a ``Domino effect''). We will introduce a generic (random walk) model for the study of cascading failures in networks, and investigate the impact of transient dynamics caused by the redistribution of loads after an initial network failure (triggering event). It is found that considering instead the stationary states, as has been done in the past, may dramatically overestimate (by 80-95%) the robustness of the network. This is due to the transient oscillations or overshooting in the loads, when the flow dynamics adjusts to the new (remaining) network structure. Consequently, additional parts of the network may be overloaded and therefore fail before the stationary state is reached. The dynamical effects are strongest on links in the neighborhood of broken links. This can cause failure waves in the network along neighboring links, while the evaluation of the stationary solution predicts a different order of link failures. March 01, 2011, 15:00 This seminar will be given on Tuesday as a Cesame seminar The main goal of the whole transcriptome analyses is to identify, characterize and catalogue all the transcripts expressed within a specific cell/tissue - at a particular stage - with the great potential to determine the correct splicing patterns and the precise structure of genes, and to quantify the differential expression of transcripts in both physiological and pathological conditions. Until 2004, hybridization and tag-based technologies, such as microarray, have allowed researchers to obtain intriguing insights into human genetics, even though microarray techniques suffer from background cross-hybridization issues and a narrow detection range, and tag-based approaches require laborious time- and cost-effective steps for the cloning of fragments prior sequencing. Recently, the introduction of massively parallel sequencing on Next Generation Sequencing (NGS) platforms has completely revolutionized the way of thinking in molecular biology. Among the different application of NGS platforms, RNA-Seq is probably one of the most complex of the various sequencing protocols developed so far. RNA-Seq has clear advantages over the existing approaches. First, RNA-Seq is not limited to the detection of known transcripts, thus allowing the identification, characterization and quantification of new splice isoforms. In addition, it permits researchers to determine the correct gene annotation, also defining the transcriptional boundaries of genes. Other advantages of RNA-Seq are the low "background signal", the absence of an upper limit for quantification and consequently, the larger dynamic range of expression levels over which transcripts can be detected. RNA-Seq data also show high levels of reproducibility. In this seminar we will illustrate some of the computational tools available for the analysis of RNA-Seq data and discuss the main challenges to face when analyzing RNA-Seq experiments. In particular we will focus on 1) Identification and quantification of transcriptional regions. 2) Identification and quantification of isoforms. 3) Detection of differential expressed events between two or more experimental conditions. 4) Connection between RNA-seq and ChIP-seq data. February 24, 2011, 14:00 The unprecedented amount of data from mobile phones creates new possibilities to analyze various facets of human behavior. Over the last few years, much attention has been devoted to studying the mobility patterns of humans. In this presentation we will focus on unusually large gatherings of people, i.e. social events. We introduce the methodology of detecting such social events in massive mobile phone data, based on a Bayesian location inference framework. More specifically, we also develop a framework for deciding who is attending an event. We demonstrate the method on a few examples. Finally we suggest some future approaches for event detection, and some possible analyses of the detected social events. February 18, 2011, 12:00 This presentation will focus on the detection and localization of communities in France based on phone calls. January 27, 2011, 11:00 Survivability in optical networks is an important issue due to the huge bandwidth offered by optical technology. Survivability means that the network has the ability to maintain an acceptable service level even after an occurrence of failures within the network. In this talk, firstly, we study and classify the various mechanisms of survivability proposed in the literature. Then we focus on p-cycles design. The major challenge of p-cycle design resides in finding an optimal set of p-cycles protecting the network for a given working capacity. January 19, 2011, 14:00 Complex networks structure is often described as composed of several groups of nodes with many links inside the groups but only a few between them. These groups are called communities and we will study the detection of such communities in evolving networks. We will present briefly some existing methods. Then we will show that classical static community algorithms fail in discovering a meaningful structure because of their instability and propose a modification to stabilize one of them. Finally we will present a new way to detect communities that are good on average and not only at one moment. December 16, 2010, 11:00 Classical spectral clustering methods are relaxations of NP-hard graph
partitioning problems. These relaxations lead to eigenvalue problems of
a graph Laplacian matrix. The relaxed solutions correspond to particular
eigenvectors which contain the grouping information. A different
formulation to spectral clustering has been introduced lately as a
primal-dual optimization setting allowing natural out-of-sample
extensions and model selection. The formulation in the dual corresponds
to an eigenvalue problem and can be interpreted as a weighted version of
kernel PCA for a particular choice of weights. December 09, 2010, 11:00 With the modern technology fast developing, most of entities can be observed by different perspectives. These multiple view information allows us to find a better pattern as long as we integrate them in an appropriate way. So clustering by integrating multi-view representations that describe the same class of entities has become a crucial issue for knowledge discovering. By extending the graph Laplacian based spectral clustering, we formulate a hybrid clustering of multiple graphs by simultaneous trace maximization. Alternating least square mechanism is adopted to obtain the optimal subspace while the effect of each graph is leveraged. A pre-processing step of dimension reduction based on multi-linear singular value decomposition (MLSVD) is employed for large-scale data. Our hybrid clustering algorithm is applied to the real application of scientific publication analysis and cross compared with the relevant hybrid clustering methods. November 24, 2010, 11:00 Given a weighted directed graph with a selected root node and a setof terminal nodes, the directed Steiner tree problem (DSTP) is tofind a directed tree of the minimal weight which is rooted in theroot node and spanning all terminal nodes. We consider the DSTPwith a constraint on the total number of arcs (hops) in the tree.This problem is known to be NP-hard, and therefore, only heuristicscan be applied in the case of large-scale instances. November 23, 2010, 11:00 We study the spreading of information in a mobile phone network by using time stamped data for an SI model. In spite of the fact that the aggregate network is a small world, the spreading is surprisingly slow. By using an appropriate set of null models we identify those correlations, which are mostly responsible for this slowing down: Topological correlations and their relation to the intensity of interactions, as well as the bursty character of activities. We give a detailed analysis of burstiness and conclude that none of the current models is able to describe the empirical findings. November 03, 2010, 11:30 We address the problem of fitting curves to data points on a manifold M. Regression serves at least two goals: (i) reducing noise on measurements and (ii) filling the gaps where data is missing. November 03, 2010, 11:00 In search engines, it is critical to be able to sort web-pages according to their relative importance. The importance of a page can be seen as the average portion of time spent in it by an infinite random walk on the web. Since the web can be modeled as a graph, the importance of a page can be measured by the PageRank of its corresponding node. Furthermore, the problem of optimizing the PageRank of a node in a directed graph when the presence of some links are uncertain is a problem that has been intensively studied lately. Here, we explain how this problem can be formulated as a Markov Decision Processes. Classical algorithms can then be used to solve this new formulation of the problem. Among these, one of them, built following the principle of Policy Iteration, seems to perform very well in practice. Yet, its theoretical behavior is unknown. In this talk, we study the complexity of this algorithm and we show that it converges in polynomial time under the natural assumption that the random walk restarts with a fixed probability called damping. We also try to extend our result to the case in which there is no damping and we give some clues towards new results in this direction. October 21, 2010, 11:00 Quadrilateral surface meshes are sometimes considered as superior to triangular meshes for finite element simulations. Discussions about if and why quadrilaterals are better than triangles are usually passionate in the finite element community. Here, we will not try to argue about that thorny question: we assume that quadrilateral meshes are useful for some people. October 14, 2010, 11:00 Complex networks appear in various contexts and are at the center of many
applied and theoretical problems. Typical examples are internet topology,
mobility graphs, peer-to-peer exchanges, web graphs, and social networks.
All these graphs have nontrivial properties in common and raise similar
questions, which makes it relevant to study them as a whole.
Most of these networks evolve with time, with the appearance and disappearance
of nodes and links. Studying these dynamics is a crucial topic, which has
however received limited attention up till now. September 08, 2010, 11:30 Problems of coordinated motion of vehicular formations have received much attention lately, especially in the areas of formation flight and so-called vehicular platoons. We investigate fundamental limitations that the use of only local feedback control has on the coherence of large formations that are subjected to stochastic disturbances. The notion of coherence translates roughly into how closely can such formations resemble a solid object, and is unrelated to the notion of string instability. This question is also of interest in the context of naturally occurring swarms, such as birds in flying formations and schools of fish. Similar dynamical phenomena occur in distributed load balancing in parallel processing and consensus-type algorithms. We examine these dynamical effects using the network topology of regular lattices, and investigate the role of topological dimension. A common phenomenon appears where a higher spatial dimension implies a more favorable scaling of coherence measures, with dimensions of 3 and 5 being necessary to achieve coherence in consensus and vehicular formations respectively with only local feedback. We show that microscopic error measures that involve only neighboring vehicles scale much more favorably, and in some cases do not suffer from this effect. This phenomenon reflects a fact that in lower spatial dimensions, local stabilizing feedbacks are relatively unable to regulate against spatially large-scale disturbances, resulting in an unregulated, "undulating" type of motion for the entire formation. We point out connections between this analysis and the statistical mechanics of harmonic solids where such phenomena have long been observed, as well as connections with the theory of resistive lattices as has been observed by others. |