se

Omar Youssef Souleimane : « Ce que j’ai entendu dans des classes de banlieue »

L'ecrivain d'origine syrienne a anime dans des colleges franciliens des ateliers organises pour lutter contre la radicalisation et promouvoir la laicite. Il temoigne.




se

Défaillances d’entreprises : n’alourdissons pas les charges

LA CHRONIQUE DE WILLIAM THAY. Le nouveau gouvernement a provoque un climat anti-riches en rompant avec la politique fiscale attractive menee ces dernieres annees.




se

La présidente du Conseil national des barreaux répond aux critiques

INTERVIEW. << J'aurais souhaite une contribution plus prospective et moins desobligeante a l'egard des barreaux de province >>, reagit M e Julie Couturier au rapport du P r Jamin.




se

Jean-Marie Le Pen hospitalisé à la suite d’un état de faiblesse général

Le fondateur du Front national, age de 96 ans et qui a mis un terme a sa vie politique, est hospitalise afin de passer plusieurs examens, par << souci de securite >>.





se

De qué se alimenta un escritor

Un escritor se tiene que alimentar. No me refiero a cómo nutrir el cuerpo. Eso es importante porque los escritores también tienen la costumbre de […]

Origen




se

Integrating Personal Web Data through Semantically Enhanced Web Portal

Currently, the World Wide Web is mostly composed of isolated and loosely connected "data islands". Connecting them together and retrieving only the information that is of interest to the user is the common Web usage process. Creating infrastructure that would support automation of that process by aggregating and integrating Web data in accordance to user's personal preferences would greatly improve today's Web usage. A significant part of Web data is available only through the login and password protected applications. As that data is very important for the usefulness of described process, proposed infrastructure needs to support authorized access to user's personal data. In this paper we propose a semantically enhanced Web portal that presents unique personalized user's entry to the domain-specific Web information. We also propose an identity management system that supports authorized access to the protected Web data. To verify the proposed solution, we have built Sweb - a semantically enhanced Web portal that uses proposed identity management system.




se

Impact of CPU-bound Processes on IP Forwarding of Linux and Windows XP

These days, commodity-off-the-shelf (COTS) hardware and software are used to build high-end and powerful workstations and servers to be deployed in today's local area networks of private homes and small- to medium-sized business. Typically, these servers are multipurpose and shared - running networking functionalities involving IP packet forwarding in addition to other CPU intensive applications. In this paper we study and investigate the impact of running CPU-bound applications on the performance of IP packet forwarding. We measure and compare the impact and performance for the two operating systems of choice for home and small-business users, namely Linux and Windows XP. The performance is studied in terms of key performance metrics which include throughput, packet loss, round-trip delay, and CPU availability. For our measurements, we consider today's typical home network hosts of modern processors and Gigabit network cards. We also consider different configuration setups and utilize open-source tools to generate relatively high traffic rates. Our empirical results show that Linux exhibits superior performance over Windows XP in terms of IP forwarding performance. Results also show that, unlike Windows XP, the IP forwarding performance of Linux is not significantly impacted by running CPU-bound applications.




se

Text Line Detection and Segmentation: Uneven Skew Angles and Hill-and-Dale Writing

In this paper a line detection and segmentation technique is presented. The proposed technique is an improved version of an older one. The experiments have been performed on the training dataset of the ICDAR 2009 handwriting segmentation contest in order to be able to compare, objectively, the performance of the two techniques. The improvement between the older and newer version is more than 24% while the average extra CPU time cost is less than 200 ms per page.




se

A New Approach to Water Flow Algorithm for Text Line Segmentation

This paper proposes a new approach to water flow algorithm for the text line segmentation. Original method assumes hypothetical water flows under a few specified angles to the document image frame from left to right and vice versa. As a result, unwetted image frames are extracted. These areas are of major importance for text line segmentation. Method modifications mean extension values of water flow angle and unwetted image frames function enlargement. Results are encouraging due to text line segmentation improvement which is the most challenging process stage in document image processing.




se

An OCR Free Method for Word Spotting in Printed Documents: the Evaluation of Different Feature Sets

An OCR free word spotting method is developed and evaluated under a strong experimental protocol. Different feature sets are evaluated under the same experimental conditions. In addition, a tuning process in the document segmentation step is proposed which provides a significant reduction in terms of processing time. For this purpose, a complete OCR-free method for word spotting in printed documents was implemented, and a document database containing document images and their corresponding ground truth text files was created. A strong experimental protocol based on 800 document images allows us to compare the results of the three feature sets used to represent the word image.




se

The Use of Latent Semantic Indexing to Mitigate OCR Effects of Related Document Images

Due to both the widespread and multipurpose use of document images and the current availability of a high number of document images repositories, robust information retrieval mechanisms and systems have been increasingly demanded. This paper presents an approach to support the automatic generation of relationships among document images by exploiting Latent Semantic Indexing (LSI) and Optical Character Recognition (OCR). We developed the LinkDI (Linking of Document Images) service, which extracts and indexes document images content, computes its latent semantics, and defines relationships among images as hyperlinks. LinkDI was experimented with document images repositories, and its performance was evaluated by comparing the quality of the relationships created among textual documents as well as among their respective document images. Considering those same document images, we ran further experiments in order to compare the performance of LinkDI when it exploits or not the LSI technique. Experimental results showed that LSI can mitigate the effects of usual OCR misrecognition, which reinforces the feasibility of LinkDI relating OCR output with high degradation.




se

The Iceberg Effect: Behind the User Interface of Mobile Collaborative Systems

Advances in mobile technologies are opening new possibilities to support collaborative activities through mobile devices. Unfortunately, mobile collaborative systems have been difficult to conceive, design and implement. These difficulties are caused in part by their unclear requirements and developers' lack of experience with this type of systems. However, several requirements involved in the collaborative back-end of these products are recurrent and should be considered in every development. This paper introduces a characterization of mobile collaboration and a framework that specifies a list of general requirements to be considered during the conception and design of a system in order to increase its probability of success. This framework was used in the development of two mobile collaborative systems, providing developers with a base of back-end requirements to aid system design and implementation. The systems were positively evaluated by their users.




se

An Empirical Study on Human and Information Technology Aspects in Collaborative Enterprise Networks

Small and Medium Enterprises (SMEs) face new challenges in the global market as customers require more complete and flexible solutions and continue to drastically reduce the number of suppliers. SMEs are trying to address these challenges through cooperation within collaborative enterprise networks (CENs). Human aspects constitute a fundamental issue in these networks as people, as opposed to organizations or Information Technology (IT) systems, cooperate. Since there is a lack of empirical studies on the role of human factors in IT-supported collaborative enterprise networks, this paper addresses the major human aspects encountered in this type of organization. These human aspects include trust issues, knowledge and know-how sharing, coordination and planning activities, and communication and mutual understanding, as well as their influence on the business processes of CENs supported by IT tools. This paper empirically proves that these aspects constitute key factors for the success or the failure of CENs. Two case studies performed on two different CENs in Switzerland are presented and the roles of human factors are identified with respect to the IT support systems. Results show that specific human factors, namely trust and communication and mutual understanding have to be well addressed in order to design and develop adequate software solutions for CENs.




se

A Petri Nets based Approach to Specify Individual and Collaborative Interaction in 3D Virtual Environments

This work describes a methodology that supports the design and implementation of software modules, which represent the individual and collaborative three-dimensional interaction process phases. The presented methodology integrates three modeling approaches: Petri Nets, a collaborative manipulation model based on the combination of single user interaction techniques taxonomy, and object-oriented programming concepts. The combination of these elements allows for the description of interaction tasks, the sequence of interaction processes being controlled by Petri Nets with the codes generated automatically. By the integration of these approaches, the present work addresses not only the entire development cycle of both individual and collaborative three-dimensional interaction, but also the reuse of developed interaction blocks in new virtual environment projects.




se

Security and Privacy Preservation for Mobile E-Learning via Digital Identity Attributes

This paper systematically discusses the security and privacy concerns for e-learning systems. A five-layer architecture of e-learning system is proposed. The security and privacy concerns are addressed respectively for five layers. This paper further examines the relationship among the security and privacy policy, the available security and privacy technology, and the degree of e-learning privacy and security. The digital identity attributes are introduced to e-learning portable devices to enhance the security and privacy of e-learning systems. This will provide significant contributions to the knowledge of e-learning security and privacy research communities and will generate more research interests.




se

The Synthesis of LSE Classifiers: From Representation to Evaluation

This work presents a first approach to the synthesis of Spanish Sign Language's (LSE) Classifier Constructions (CCs). All current attempts at the automatic synthesis of LSE simply create the animations corresponding to sequences of signs. This work, however, includes the synthesis of the LSE classification phenomena, defining more complex elements than simple signs, such as Classifier Predicates, Inflective CCs and Affixal classifiers. The intelligibility of our synthetic messages was evaluated by LSE natives, who reported a recognition rate of 93% correct answers.




se

On Compound Purposes and Compound Reasons for Enabling Privacy

This paper puts forward a verification method for compound purposes and compound reasons to be used during purpose limitation.

When it is absolutely necessary to collect privacy related information, it is essential that privacy enhancing technologies (PETs) protect access to data - in general accomplished by using the concept of purposes bound to data. Compound purposes and reasons are an enhancement of purposes used during purpose limitation and binding and are more expressive than purposes in their general form. Data users specify their access needs by making use of compound reasons which are defined in terms of (compound) purposes. Purposes are organised in a lattice with purposes near the greatest lower bound (GLB) considered weak (less specific) and purposes near the least upper bound (LUB) considered strong (most specific).

Access is granted based on the verification of the statement of intent (from the data user) against the compound purpose bound to the data; however, because purposes are in a lattice, the data user is not limited to a statement of intent that matches the purposes bound to the data exactly - the statement can be a true reflection of their intent with the data. Hence, the verification of compound reasons against compound purposes cannot be accomplished by current published verification algorithms.

Before presenting the verification method, compound purposes and reasons, as well as the structures used to represent them, and the operators that are used to define compounds is presented. Finally, some thoughts on implementation are provided.




se

Pragmatic Knowledge Services

Knowledge, innovations and their implementation in effective practices are essential for development in all fields of societal action, e.g. policy, business, health, education, and everyday life. However, managing the interrelations between knowledge, innovation and practice is complicated. Facilitation by suitable knowledge services is needed. This paper explores the theory of converging knowledge, innovation, and practice, discusses some advances in information systems development, and identifies general requirements for pragmatic knowledge services. A trialogical approach to knowledge creation and learning is adopted as a viable theoretical basis. Also three examples of novel knowledge services Opasnet, Innovillage, and Knowledge Practices Environment (KPE), are presented. Eventually, it is concluded that pragmatic knowledge services, as hybrid systems of information technology and its users, are not only means for creation of practical knowledge, but vehicles of a cultural change from individualistic perceptions of knowledge work towards mediated collaboration.




se

IDEA: A Framework for a Knowledge-based Enterprise 2.0

This paper looks at the convergence of knowledge management and Enterprise 2.0 and describes the possibilities for an over-arching exchange and transfer of knowledge in Enterprise 2.0. This will be underlined by the presentation of the concrete example of T-System Multimedia Solutions (MMS), which describes the establishment of a new enterprise division "IG eHealth". This is typified by the decentralised development of common ideas, collaboration and the assistance available to performing responsibilities as provided by Enterprise 2.0 tools. Taking this archetypal example and the derived abstraction of the problem regarding the collaboration of knowledge workers as the basis, a regulatory framework will be developed for knowledge management to serve as a template for the systemisation and definition of specific Enterprise 2.0 activities. The paper will conclude by stating factors of success and supporting Enterprise 2.0 activities, which will facilitate the establishment of a practical knowledge management system for the optimisation of knowledge transfer.




se

Enterprise Microblogging for Advanced Knowledge Sharing: The References@BT Case Study

Siemens is well known for ambitious efforts in knowledge management, providing a series of innovative tools and applications within the intranet. References@BT is such a web-based application with currently more than 7,300 registered users from more than 70 countries. Its goal is to support the sharing of knowledge, experiences and best-practices globally within the Building Technologies division. Launched in 2005, References@BT features structured knowledge references, discussion forums, and a basic social networking service. In response to use demand, a new microblogging service, tightly integrated into References@BT, was implemented in March 2009. More than 500 authors have created around 2,600 microblog postings since then. Following a brief introduction into the community platform References@BT, we comprehensively describe the motivation, experiences and advantages for an organization in providing internal microblogging services. We provide detailed microblog usage statistics, analyzing the top ten users regarding postings and followers as well as the top ten topics. In doing so, we aim to shed light on microblogging usage and adoption within a globally distributed organization.




se

Bio-Inspired Mechanisms for Coordinating Multiple Instances of a Service Feature in Dynamic Software Product Lines

One of the challenges in Dynamic Software Product Line (DSPL) is how to support the coordination of multiple instances of a service feature. In particular, there is a need for a decentralized decision-making capability that will be able to seamlessly integrate new instances of a service feature without an omniscient central controller. Because of the need for decentralization, we are investigating principles from self-organization in biological organisms. As an initial proof of concept, we have applied three bio-inspired techniques to a simple smart home scenario: quorum sensing based service activation, a firefly algorithm for synchronization, and a gossiping (epidemic) protocol for information dissemination. In this paper, we first explain why we selected those techniques using a set of motivating scenarios of a smart home and then describe our experiences in adopting them.




se

QoS-based Approach for Dynamic Web Service Composition

Web Services have become a standard for integration of systems in distributed environments. By using a set of open interoperability standards, they allow computer-computer interaction, regardless the programming languages and operating systems used. The Semantic Web Services, by its turn, make use of ontologies to describe their functionality in a more structural manner, allowing computers to reason about the information required and provided by them. Such a description also allows dynamic composition of several Web Services, when only one is not able to provide the desired functionality. There are scenarios, however, in which only the functional correctness is not enough to fulfill the user requirements, and a minimum level of quality should be guaranteed by their providers. In this context, this work presents an approach for dynamic Web Service composition that takes into account the composition overall quality. The proposed approach relies on a heuristics to efficiently perform the composition. In order to show the feasibility of the proposed approach, a Web Service composition application prototype was developed and experimented with public test sets, along with another approach that does not consider quality in the composition process. The results have shown that the proposed approach in general finds compositions with more quality, within a reasonable processing time.




se

An Aspect-Oriented Framework for Weaving Domain-Specific Concerns into Component-Based Systems

Software components are used in various application domains, and many component models and frameworks have been proposed to fulfill domain-specific requirements. The general trend followed by these approaches is to provide ad-hoc models and tools for capturing these requirements and for implementing their support within dedicated runtime platforms, limited to features of the targeted domain. The challenge is then to propose more flexible solutions, where components reuse is domain agnostic. In this article, we present a framework supporting compositional construction and development of applications that must meet various extra-functional/domain-specific requirements. The key points of our contribution are: i) We target development of component-oriented applications where extra-functional requirements are expressed as annotations on the units of composition in the application architecture. ii) These annotations are implemented as open and extensible component-based containers, achieving full separation of functional and extra-functional concerns. iii) Finally, the full machinery is implemented using the Aspect-Oriented Programming paradigm. We validate our approach with two case studies: the first is related to real-time and embedded applications, while the




se

Context-Aware Composition and Adaptation based on Model Transformation

Using pre-existing software components (COTS) to develop software systems requires the composition and adaptation of the component interfaces to solve mismatch problems. These mismatches may appear at different interoperability levels (signature, behavioural, quality of service and semantic). In this article, we define an approach which supports composition and adaptation of software components based on model transformation by taking into account the four levels. Signature and behavioural levels are addressed by means of transition systems. Context-awareness and semanticbased techniques are used to tackle quality of service and semantic, respectively, but also both consider the signature level. We have implemented and validated our proposal for the design and application of realistic and complex systems. Here, we illustrate the need to support the variability of the adaptation process in a context-aware pervasive system through a real-world case study, where software components are implemented using Windows Workflow Foundation (WF). We apply our model transformation process to extract transition systems (CA-STS specifications) from WF components. These CA-STSs are used to tackle the composition and adaptation. Then, we generate a CASTS adaptor specification, which is transformed into its corresponding WF adaptor component with the purpose of interacting with all the WF components of the system, thereby avoiding mismatch problems.




se

Software Components, Architectures and Reuse




se

Improving Security Levels of IEEE802.16e Authentication by Involving Diffie-Hellman PKDS

Recently, IEEE 802.16 Worldwide Interoperability for Microwave Access (WiMAX for short) has provided us with low-cost, high efficiency and high bandwidth network services. However, as with the WiFi, the radio wave transmission also makes the WiMAX face the wireless transmission security problem. To solve this problem, the IEEE802.16Std during its development stage defines the Privacy Key Management (PKM for short) authentication process which offers a one-way authentication. However, using a one-way authentication, an SS may connect to a fake BS. Mutual authentication, like that developed for PKMv2, can avoid this problem. Therefore, in this paper, we propose an authentication key management approach, called Diffie-Hellman-PKDS-based authentication method (DiHam for short), which employs a secret door asymmetric one-way function, Public Key Distribution System (PKDS for short), to improve current security level of facility authentication between WiMAX's BS and SS. We further integrate the PKMv1 and the DiHam into a system, called PKM-DiHam (P-DiHam for short), in which the PKMv1 acts as the authentication process, and the DiHam is responsible for key management and delivery. By transmitting securely protected and well-defined parameters for SS and BS, the two stations can mutually authenticate each other. Messages including those conveying user data and authentication parameters can be then more securely delivered.




se

Hierarchical Graph-Grammar Model for Secure and Efficient Handwritten Signatures Classification

One important subject associated with personal authentication capabilities is the analysis of handwritten signatures. Among the many known techniques, algorithms based on linguistic formalisms are also possible. However, such techniques require a number of algorithms for intelligent image analysis to be applied, allowing the development of new solutions in the field of personal authentication and building modern security systems based on the advanced recognition of such patterns. The article presents the approach based on the usage of syntactic methods for the static analysis of handwritten signatures. The graph linguistic formalisms applied, such as the IE graph and ETPL(k) grammar, are characterised by considerable descriptive strength and a polynomial membership problem of the syntactic analysis. For the purposes of representing the analysed handwritten signatures, new hierarchical (two-layer) HIE graph structures based on IE graphs have been defined. The two-layer graph description makes it possible to take into consideration both local and global features of the signature. The usage of attributed graphs enables the storage of additional semantic information describing the properties of individual signature strokes. The verification and recognition of a signature consists in analysing the affiliation of its graph description to the language describing the specimen database. Initial assessments display a precision of the method at a average level of under 75%.




se

Cost-Sensitive Spam Detection Using Parameters Optimization and Feature Selection

E-mail spam is no more garbage but risk since it recently includes virus attachments and spyware agents which make the recipients' system ruined, therefore, there is an emerging need for spam detection. Many spam detection techniques based on machine learning techniques have been proposed. As the amount of spam has been increased tremendously using bulk mailing tools, spam detection techniques should counteract with it. To cope with this, parameters optimization and feature selection have been used to reduce processing overheads while guaranteeing high detection rates. However, previous approaches have not taken into account feature variable importance and optimal number of features. Moreover, to the best of our knowledge, there is no approach which uses both parameters optimization and feature selection together for spam detection. In this paper, we propose a spam detection model enabling both parameters optimization and optimal feature selection; we optimize two parameters of detection models using Random Forests (RF) so as to maximize the detection rates. We provide the variable importance of each feature so that it is easy to eliminate the irrelevant features. Furthermore, we decide an optimal number of selected features using two methods; (i) only one parameters optimization during overall feature selection and (ii) parameters optimization in every feature elimination phase. Finally, we evaluate our spam detection model with cost-sensitive measures to avoid misclassification of legitimate messages, since the cost of classifying a legitimate message as a spam far outweighs the cost of classifying a spam as a legitimate message. We perform experiments on Spambase dataset and show the feasibility of our approaches.




se

Service Oriented Multimedia Delivery System in Pervasive Environments

Service composition is an effective approach for large-scale multimedia delivery. In previous works, user requirement is represented as one fixed functional path which is composed of several functional components in a certain order. Actually, there may be several functional paths (deliver different quality level multimedia data, e.g., image pixel, frame rate) that can meet one request. And due to the diversity of devices and connections in pervasive environment, system should choose a suitable media quality delivery path in accordance with context, instead of one fixed functional path. This paper presents a deep study of multimedia delivery problem and proposes an on-line algorithm LDPath and an off-line centralized algorithm LD/RPath respectively. LDPath aims at delivering multimedia data to end user with lowest delay by choosing services to build delivery paths hop-by-hop, which is adapted to the unstable open environment. And LD/RPath is developed for a relatively stable environment, which generates delivery paths according to the trade-off between delay and reliability metrics, because the service reliability is also an important fact in such scenario. Experimental results show that both algorithms have good performance with low overhead to the system.




se

Semantic Web: Theory and Applicationsns




se

Towards Classification of Web Ontologies for the Emerging Semantic Web

The massive growth in ontology development has opened new research challenges such as ontology management, search and retrieval for the entire semantic web community. These results in many recent developments, like OntoKhoj, Swoogle, OntoSearch2, that facilitate tasks user have to perform. These semantic web portals mainly treat ontologies as plain texts and use the traditional text classification algorithms for classifying ontologies in directories and assigning predefined labels rather than using the semantic knowledge hidden within the ontologies. These approaches suffer from many types of classification problems and lack of accuracy, especially in the case of overlapping ontologies that share common vocabularies. In this paper, we define an ontology classification problem and categorize it into many sub-problems. We present a new ontological methodology for the classification of web ontologies, which has been guided by the requirements of the emerging Semantic Web applications and by the lessons learnt from previous systems. The proposed framework, OntClassifire, is tested on 34 ontologies with a certain degree of overlapping domain, and effectiveness of the ontological mechanism is verified. It benefits the construction, maintenance or expansion of ontology directories on the semantic web that help to focus on the crawling and improving the quality of search for the software agents and people. We conclude that the use of a context specific knowledge hidden in the structure of ontologies gives more accurate results for the ontology classification.




se

A Semantic Wiki Based on Spatial Hypertext

Spatial Hypertext Wiki (ShyWiki) is a wiki which represents knowledge using notes that are spatially distributed in wiki pages and have visual characteristics such as colour, size, or font type. The use of spatial and visual characteristics in wikis is important to improve human comprehension, creation and organization of knowledge. Another important capability in wikis is to allow machines to process knowledge. Wikis that formally structure knowledge for this purpose are called semantic wikis. This paper describes how ShyWiki can make use of spatial hypertext in order to be a semantic wiki. ShyWiki can represent knowledge at different levels of formality. Users of ShyWiki can annotate the content and represent semantic relations without being experts of semantic web data description languages. The spatial hypertext features make it suitable for users to represent unstructured knowledge and implicit graphic relations among concepts. In addition, semantic web and spatial hypertext features are combined to represent structured knowledge. The semantic web features of ShyWiki improve navigation and publish the wiki knowledge as RDF resources, including the implicit relations that are analyzed using a spatial parser.




se

A Ranking Tool Exploiting Semantic Descriptions for the Comparison of EQF-based Qualifications

Nowadays, one of the main issues discussed at the Community level is represented by the mobility of students and workers across Europe. During the last years, in order to deal with the above picture, several initiatives have been carried out: one of them is the definition of the European Qualification Framework (EQF), a common architecture for the description of qualifications. At the same time, several research activities were established with the aim of finding how semantic technologies could be exploited for qualifications comparison in the field of human resources acquisition. In this paper, the EQF specifications are taken into account and they are applied in a practical scenario to develop a ranking algorithm for the comparison of qualifications expressed in terms of knowledge, skill and competence concepts, potentially aimed at supporting European employers during the recruiting phase.




se

Ontology-based User Interface Development: User Experience Elements Pattern

The user experience of any software or website consists of elements from the conceptual to the concrete level. These elements of user experience assist in the design and development of user interfaces. On the other hand, ontologies provide a framework for computable representation of user interface elements and underlying data. This paper discusses strategies of introducing ontologies at different user interface layers adapted from user experience elements. These layers range from abstract levels (e.g. User needs/Application Objectives) to concrete levels (e.g. Application User Interface) in terms of data representation. The proposed ontological framework enables device independent, semi-automated GUI construction which we will demonstrate at a personal information management example.




se

Ontology-based Competency Management: the Case Study of the Mihajlo Pupin Institute

Semantic-based technologies have been steadily increasing their relevance in recent years in both the research world and business world. Considering this, the present article discusses the process of design and implementation of a competency management system in information and communication technologies domain utilizing the latest Semantic Web tools and technologies including D2RQ server, TopBraid Composer, OWL 2, SPARQL, SPARQL Rules and common human resources related public vocabularies. In particular, the paper discusses the process of building individual and enterprise competence models in a form of ontology database, as well as different ways of meaningful search and retrieval of expertise data on the Semantic Web. The ontological knowledge base aims at storing the extracted and integrated competences from structured, as well as unstructured sources. By using the illustrative case study of deployment of such a system in the Human Resources sector at the Mihajlo Pupin Institute, this paper shows an example of new approaches to data integration and information management. The proposed approach extends the functionalities of existing enterprise information systems and offers possibilities for development of future Internet services. This allows organizations to express their core competences and talents in a standardized, machine processable and understandable format, and hence, facilitates their integration in the European Research Area and beyond.




se

An Ontology based Agent Generation for Information Retrieval on Cloud Environment

Retrieving information or discovering knowledge from a well organized data center in general is requested to be familiar with its schema, structure, and architecture, which against the inherent concept and characteristics of cloud environment. An effective approach to retrieve desired information or to extract useful knowledge is an important issue in the emerging information/knowledge cloud. In this paper, we propose an ontology-based agent generation framework for information retrieval in a flexible, transparent, and easy way on cloud environment. While user submitting a flat-text based request for retrieving information on a cloud environment, the request will be automatically deduced by a Reasoning Agent (RA) based on predefined ontology and reasoning rule, and then be translated to a Mobile Information Retrieving Agent Description File (MIRADF) that is formatted in a proposed Mobile Agent Description Language (MADF). A generating agent, named MIRA-GA, is also implemented to generate a MIRA according to the MIRADF. We also design and implement a prototype to integrate these agents and show an interesting example to demonstrate the feasibility of the architecture.




se

ORPMS: An Ontology-based Real-time Project Monitoring System in the Cloud

Project monitoring plays a crucial role in project management, which is a part of every stage of a project's life-cycle. Nevertheless, along with the increasing ratio of outsourcing in many companies' strategic plans, project monitoring has been challenged by geographically dispersed project teams and culturally diverse team members. Furthermore, because of the lack of a uniform standard, data exchange between various project monitoring software becomes an impossible mission. These factors together lead to the issue of ambiguity in project monitoring processes. Ontology is a form of knowledge representation with the purpose of disambiguation. Consequently, in this paper, we propose the framework of an ontology-based real-time project monitoring system (ORPSM), in order to, by means of ontologies, solve the ambiguity issue in project monitoring processes caused by multiple factors. The framework incorporates a series of ontologies for knowledge capture, storage, sharing and term disambiguation in project monitoring processes, and a series of metrics for assisting management of project organizations to better monitor projects. We propose to configure the ORPMS framework in a cloud environment, aiming at providing the project monitoring service to geographically distributed and dynamic project members with great flexibility, scalability and security. A case study is conducted on a prototype of the ORPMS in order to evaluate the framework.




se

Cooperation as a Service in VANETs

Vehicular Networks, including Vehicular Adhoc Networks (VANETs) and Vehicular Sensor Networks (VSNs), stimulate a brand new variety of services, ranging from driver safety services, traffic information and warnings regarding traffic jams and accidents, to providing weather or road condition, parking availability, and advertisement. 3G networks and sophisticated Intelligent Transportation Systems (ITS), including deploying costly roadside base stations, can indeed be used to offer such services, but these come with a cost, both at network and hardware levels. In this paper we introduce Cooperation as a service (CaaS): A novel architecture that will allow providing a set of services for free and without any additional infrastructure, by taking advantage of Vehicle-to-Vehicle communications. CaaS uses a hybrid publish/subscribe mechanism where the driver (or subscriber) expresses his interests regarding a service (or a set of services) and where cars having subscribed to the same service will cooperate to provide the subscriber with the necessary information regarding the service he subscribed to, by publishing this information in the network. CaaS structures the network into clusters, and uses Content Based Routing (CBR) for intra-cluster communications and geographic routing for inter-cluster communications.




se

Moralisation : Les annonces de Bayrou vont dans le bon sens

Malgré un télescopage plus que dommageable avec l’affaire Ferrand – il aurait déjà dû démissionner – François Bayrou a annoncé un train de mesures visant à moraliser visant à encadrer les élus. Et bien celles-ci vont dans le bon sens. Bien sûr, on pourra...




se

Réseaux Sociaux… entre hystérisation et égotisme

Si la question de l’hystérisation des réseaux sociaux ne date pas des Gilets Jaunes, nulle doute qu’elle revient sur le devant de la scène. Je ne parle même pas des théories du complot lié à ce mouvement. Pour ma part, les choses sont claires, il est...




se

Strasbourg et la récurrence complotiste des Réseaux Sociaux.

Comme on devait s’y attendre, l’attentat de Strasbourg n’a pas manqué, s’il en était encore besoin, de charrier son lot de rumeurs complotistes. Un regard rapide sur les réseaux sociaux de toute nature suffit à se rendre compte de ce triste phénomène....




se

On reparle des classes populaires, des péri-urbains, partis politiques... (billet de Mai 2011 )

L'actualité aidant, une personne sur twitter a exhumé un vieux billet que j'avais écrit en Mai 2011. A l'époque, je parlais du Parti Socialiste... sans fausse modestie, j'avais vu plutôt juste... Aujourd'hui, on peut déjà parler d'un autre "parti" mais...




se

Le Référendum d'Initiative Citoyenne ou le sexe des anges...

Mine de rien les Gilets jaunes, quoiqu'on en pense, on réussit à faire bouger les lignes sur des éléments qu'on nous disait impossible jusqu'à présent. Au delà de certaines applications qui peuvent être laissées au débat, en matière de pouvoir d'achat,...




se

Intelligence Artificielle : vers le grand déclassement des Classes Moyennes ?

Depuis quelques années, la théorie du grand remplacement, popularisée par Michel Houellbecq dans Soumissions ou par un Eric Zemmour, a fait son chemin dans les arcanes les moins visibles du Net. Pourtant, le danger n’est pas là, loin s’en faut, il est...




se

Emmanuel Macron et les Réseaux Sociaux

C’est entendu pour Emmanuel Macron, les violences auraient pour cause les Réseaux Sociaux. Je cite « il y a un changement anthropologique de nos sociétés qui vient des réseaux sociaux ». J’avoue que ce type de raisonnement aussi simplistes me laisse pantois....




se

Tensions à l'Université Lyon-3 : la classe politique condamne les attaques contre Yaël Braun-Pivet

Tensions à l'Université Lyon-3 : la classe politique condamne les attaques contre Yaël Braun-Pivet




se

Avec l'Arizona, Donald Trump a désormais remporté la totalité des sept États clés

Avec l'Arizona, Donald Trump a désormais remporté la totalité des sept États clés




se

Trump et Poutine auraient discuté de la guerre en Ukraine, "une pure fiction", selon le Kremlin

Trump et Poutine auraient discuté de la guerre en Ukraine, "une pure fiction", selon le Kremlin




se

No Comment : les mariachis se réunissent à Mexico pour battre le record du monde de chant

No Comment : les mariachis se réunissent à Mexico pour battre le record du monde de chant