internet

Left out of the digital revolution?

Dominique Pasquier, Télécom ParisTech, Institut Mines-Télécom (IMT)

This text is published as part of the “Digital Society” column written by researchers from the Economic and Social Sciences department at Télécom ParisTech, members of the Interdisciplinary Institute for Innovation (CNRS).

[divider style=”dotted” top=”20″ bottom=”20″]

[dropcap]S[/dropcap]ome revolutions are silent ones. Internet integration in working-class areas is one such example.

Where do we stand in terms of the “digital divide”?

In the early 2000s there was a lot of talk about the “digital divide” focusing on inequality in terms of both access and uses. Annual reports by CREDOC have shown that over the past ten years working-class categories have started to catch up in terms of internet access: in France, between 2006 and 2017, the proportion of employees with Internet access at home increased from 51% to 93%, while for blue collars it rose from 38% to 83% (CREDOC 2017 : 48).

Age, rather than income or level of education, is now the most determining factor (8 out of 10 of those who never use the internet are 60 or older). But these same reports show that while the divide is closing in terms of internet access, internet use among working and lower classes remains less varied and frequent than among the middle and upper classes. Individuals without college degrees find it harder to adjust to paperless administrative services, do less research, buy less online, and only very rarely produce content. In short, there appears to be a kind of “working classes’ internet” that is less creative, less daring and less useful to a certain extent.

Shifting the focus

Perhaps the question should be looked at from a different angle. These statistical surveys are based on tallying up and lamenting shortcomings in comparison to the most innovative internet practices, meaning those of young, educated urban populations. The issue could be examined from a different perspective by starting out with the idea that the internet practices favored by the working and lower classes make sense in light of their everyday needs and that they are important indicators for understanding their relationship with the world and possible transformations in this relationship.

As Jacques Rancière explained in his analysis of written productions of workers in the 19th century, it is a matter of setting the equality of intelligences as a starting point for discussing and understanding how “a common language appropriated by others” can be re-appropriated by those for whom it was not intended (Rancière 2009 : 152).

This shifted focus would make it possible to perceive uses that may not stand out as remarkable, but which have dramatically transformed the relationship with knowledge and learning among those with low levels of education. Examples of such uses include looking up words used by doctors or found in headings of children’s homework. For sophisticated internet users, these uses may seem rather basic, but in the meantime they bring about profound changes by making relationships with experts less asymmetrical and by reducing the phenomena of “deferential attitude” among the working classes that Annette Lareau analyzes in her excellent book, Unequal Childhoods (2011).

Online research: learning and shopping

Employees with lower-paying jobs who do not use digital technology for work also spend a great deal of time online to inform themselves about their profession or their rights: the success of websites for caregivers is a good example. Childcare workers discuss their views on bringing up children and hospital service workers talk about their relationships with patients. It is also important to note the ways in which tutorials have revived interest in expertise traditionally held by the working class: novel cooking ingredients, new gardening or DIY methods and never-seen-before knitting patterns have made their way into homes.

Working-class families thus use the internet to learn, but they also use it make purchases. For people who live in rural or semi-rural areas, being able to access goods that had previously been impossible to find nearby in just a few clicks would appear to represent a huge opportunity. But in reality, it is a bit more complicated. Online marketplaces are less appreciated for their choice as they are for the savings they provide through browsing for special offers. Looking for deals is the main source of motivation, along with the fact that this makes it possible to manage supplies by purchasing in batches. At the same time, these savings are guilty of playing a role in weakening local business – or what remains of local business anyway.

In communities where people know each other well, shopkeepers are also neighbors – and sometimes friends – and betrayal can leave both sides feeling bitter. On the other hand, marketplaces for secondhand goods, such as Le Bon Coin which has recruited a significant number of rural and working-class customers, are described as virtuous markets: they allow people to enjoy leisurely browsing by geographic location (this has become a new source of gossip!) and provide an alternative to throwing away used goods. These marketplaces also give people the opportunity to earn a little money and help those who buy goods maintain a sense of pride, since they are able to furnish their homes or buy clothes at a low cost without resorting to donation systems. Online shopping has therefore led to paradoxical changes in relationships with the local community, as it has destroyed certain ties but created others.

Using the internet for reading and communication

The internet can be described as a relationship with writing – the mark of those who have created it. This can be difficult for individuals with low levels of education who do not write much in their work. Email, which requires standardized writing, is largely underused in these working-class families: it is only used to communicate with e-commerce websites or administrative services, and is often related to a problem in the latter case.

This is also because email is based on an interpersonal, asynchronous communication rationale that goes against the standards of face-to-face relationships and group discussion that are so prevalent in working-class communities. Facebook has fared much better: it allows users to exchange content by sharing links, is in keeping with a group discussion system and does not require formal writing of any kind. This social network appears to be a good way to stay in touch with extended family and close friends, while seeking consensus on shared values. It is a self-segregating network which is not particularly open to other social worlds.

While the internet has kept many of its promises in terms of our relationship with knowledge, it clearly has not managed to break down the barriers between different social worlds.

[divider style=”dotted” top=”20″ bottom=”20″]

Dominique Pasquier, sociologist, research director at CNRS and author of L’Internet des familles modestes. Enquête dans la France rurale (The Internet of Low-Income Families. A Survey in Rural France) Paris, Presses des Mines, 2018.

Dominique Pasquier, sociologist, research director at CNRS, member of the Interdisciplinary Institute for Innovation (i3), Télécom ParisTech, Institut Mines-Télécom.

The original version of this article was published in French on The Conversation France.

Also read on I’MTech

IIoT

What is the Industrial Internet of Things (IIoT)?

Industry and civil society do not share the same expectations when it comes to connected objects. The Internet of Things (IoT) must therefore adapt to meet industrial demands. These specific adaptations have led to the emergence of a new field: IIoT, or the Industrial Internet of Things. Nicolas Montavont, a researcher at IMT Atlantique, describes the industrial stakes that justify the specific nature of the IIoT and the challenges currently facing the scientific community.

 

What does the IIoT look like in specific terms?

Nicolas Montavont: One of the easiest examples to present and understand is the way production lines are monitored. Sensors ensure that a product is manufactured under good conditions, by controlling what travels down the conveyor belt and by measuring the temperature, humidity or luminosity for the specific work environment. Actuators can then respond to the data received, for example by reconfiguring a production line based on the environment or context, allowing a machine to perform a different task.

How does the IIoT benefit companies?

NM: There are benefits in every area: production times, line performance, cost reduction, etc. One major benefit is increased flexibility thanks to a more autonomous system. Production lines can operate and adapt with fewer human interventions. Staff can therefore transition from a role of handling and management to supervision and control. This change especially benefits small businesses. Today, production is very focused on large volumes. Increased flexibility and autonomy let companies find more cost-effective ways of manufacturing small quantities.

What justifies referring to IIoT as a separate field, distinct from the mainstream IoT?

NM: Mainstream IoT technologies are not designed to meet industry requirements. In general, when IoT is used for applications, the performance requirements are not very high. Communicating objects are used to send non-critical data packets without strict time constraints. The opposite is true for industry demands, which require object networks that send important data with the lowest possible latency. Therefore, specific IoT standards must be developed for the industrial sector, hence the name IIoT. For example, companies do not want to be limited by propriety standards, and so they want to push the Internet to become the network that supports their architectures.

Why do companies have more performance constraints for their networks of communicating objects?

NM: One scenario that clearly represents industrial constraints is the one we selected for the SCHEIF project, in the context of the German-French Academy for the Industry of the Future (GFA). We initiated a collaboration with the Technische Universität of Munich (TUM) on the quality of the network and data in an industrial environment. We started with a scenario featuring a set of robots that move autonomously through a work environment. They can accomplish specific tasks and can also detect and adapt to environmental changes. For example, if a person walks through the area, they must not be hit by the robots. This scenario includes a major safety aspect, which demands an efficient network, low latency, good-quality data communications and the effective assessment of the state of the environment.

What are the scientific challenges of this type of example?

NM: First of all, locating robots indoors in real time presents a challenge. Technologies exist but are not yet perfect and do not offer sufficient performance levels. Secondly, we need to make sure the robots exchange the monitoring data in an appropriate manner, by prioritizing the information. The main problem is, “who needs to send what, and when?” We are working on how to schedule the communications and represent the robots’ knowledge of their environment. We also have energy consumption constraints, first in terms of hardware and the network. Finally, there is a significant cybersecurity aspect, which has become a major focus area for the scientific community.

digital simulation

What is digital simulation?

Digital simulation has become an almost mandatory step in developing new products. But what does “simulating behavior” or “modeling an event” really mean? Marius Preda, a computer science researcher at Télécom SudParis, explains what’s hiding behind these common industry expressions.

 

What is digital simulation used for?

Marius Preda: Its main goal is to reduce prototyping costs for manufacturers. Instead of testing a product with real prototypes, which are expensive, companies use fully digital twins of these prototypes. These virtual twins take the form of a 3D model that has all the same attributes as the real product– colors, dimensions, visual aspect– and most importantly, in which a great quantity of metadata is injected, such as physical properties of the materials. This makes it possible to create a simulation that is very close to reality. The obvious advantage is that if the product isn’t right, the metadata can simply be changed, or the attributes of the digital twin can be directly modified. With a real prototype, it would have to be entirely remade.

What can be simulated?

MP: The main focus is on production. Companies make simulations in order to accurately measure all the parameters of a part and obtain production specifications. That explains a high percentage of uses of digital simulation. After that, there are significant concerns about aging. Physical laws that determine how materials wear are well-known, so companies inject them into digital models in order to simulate how a part will wear as it is used. One of the new applications is predictive maintenance. Simulations can be used to predict breakage or faults in order to determine the optimal moment a part should be repaired or replaced. All that relates to products but there are also simulations of whole factories to simulate their operations, and simulations of the human body.

Read on IMTech:  A digital twin of the aorta to help prevent aneurysm rupture

How is a digital simulation created?

MP: The first step is defining the goal of the simulation. Taking a car, for example, if the goal is to study how the car body deforms during an impact, the modeling will be different from if the goal were to analyze visual and sound comfort inside the passenger compartment. So modeling is carried out based on what the aim is: automobile manufacturers don’t create a 3D model with the idea that they’ll be able to use it for all simulations. The 3D form may be the same, but what’s important are the physical properties that will be included within the model. For crash test simulations, properties related to the way materials deform are injected into the model in the form of equations that govern their behavior. For sound comfort, the laws of reflectivity and sound propagation are included.

What form do simulations take?

MP: Virtual reality is often presented as something new, but manufacturers have been using it for years for simulations! In the past, they would create 3D environments called “caves,” which where rooms in which different parts of a car – to continue with our automobile example – were projected on the walls. Today, virtual reality headsets make it possible to save space and put more people in the same virtual environment. But beyond this highly visual form of simulation, what industry professionals are really interested in is the model and the results behind it. What matters isn’t really seeing how a car deforms in an accident, but knowing by how many centimeters the engine block penetrates into the passenger compartment. And sometimes, there isn’t even a visual: the simulation takes the form of a curve on a graph showing how material deformation depends on the speed of the car.

What sectors use digital simulations the most?  

MP: I talk about the automobile industry a lot since it’s one of the first to have used digital simulations. Architects were also among the first to use 3D to visualize models. And factories and relatively complex industrial facilities rely on simulation too. Among other things, it allows them to analyze the piping systems behind the walls. It’s a way to access information more easily than with plans. On the other hand, there are sectors, such as construction and civil engineering, where simulation is under-utilized and plans still play a central role.

What are some major ways digital simulation could evolve in the near future?

MP: In my opinion, interaction between humans and 3D models represents a big challenge. New devices like virtual reality glasses are being used, but the way people interact with the model remains unnatural. Yes, from within a virtual space, users can change how rooms are arranged with a wave of the hand. But if they want to change the physical parameters behind a material’s behavior, they still have to use a computer to introduce raw data in coded form. It would be a major advance to be able to directly change these metadata from within the virtual environment.

 

healthcare

AI in healthcare for the benefit of individuals and society?

Article written by Christian Roux (Director of Research and Innovation at IMT), Patrick Duvaut (Director of Innovation at IMT), and Eric Vibert (professor at Université Paris-Sud/Université Paris Saclay, and surgeon at Hôpital Paul Brousse (AP-HP) in Villejuif).

[divider style=”normal” top=”20″ bottom=”20″]

How can artificial intelligence be built in such a way that it is humanistic, explainable and ethical? This question is central to discussions about the future of AI technology in healthcare. This technology must provide humans with additional skills without replacing healthcare professionals. Focusing on the concept of “sustainable digital technology,” this article presents the key ideas explained by the authors of a long format report published in French in the Télécom Alumni journal. 

 

Artificial intelligence (AI) is still in its early stages. Performing single tasks and requiring huge numbers of examples, it lacks a conscience, empathy and common sense. In the medical field, the relationship between patient and augmented caregiver, or virtual medical assistant, is a key issue. In AI, what is needed is a personalized, social learning behavior between machines and patients. There is also a technological and scientific limitation: the lack of explainability in AI. Methods such as deep learning act like black boxes. A single verdict resulting from a process that can be described as “input big data, output results” is not sufficient for practitioners or patients. For the former it is an issue of taking responsibility for supporting, understanding and controlling the patient treatment pathway, while for the latter, it is one of adhering to a treatment and therapy approach, crucial aspects of patients’ involvement in their own care.

However, the greatest challenges facing AI are not technological, but are connected to its use, governance, ethics, etc. Going from “cognitive,” and “social” technologies to decision-making technologies is a major step. It means completely handing over responsibility to digital technology. The methodological bases required for such a transition exist, but putting them to use would have a significant impact on the nature of the coevolution of humans and AI. As far as doctors are concerned, the January 2018 report by the French Medical Council (CNOM) is quite clear: practitioners wish to keep control over decision-making. “Machines must serve man, not control him,” states the report. Machines must be restricted to augmenting decision-making or diagnoses, like IBM Watson. As philosopher and essayist Miguel Benasayag astutely points out, “AI does not ask questions, Man has to ask them.”

Leaving humans in charge of decision-making augmented by AI has become an even more crucial issue in recent years. Since 2016, society has been facing its most serious crisis of confidence in social media and digital service platforms since the dawn of the digital age, with the creation of the World Wide Web in 1990. After nearly 30 years of existence, digital society is going through a paradigm shift. Facing pressure from citizens driven by a crisis of confidence, it is arming itself with new tools, such as the European Data Protection Regulation (GDPR). The age of alienating digital technology, which was free to act as a cognitive, social and political colonizer for three decades, is being replaced by “sustainable digital technology” that puts citizens at the center of the cyber-sphere by giving them control over their data and their digital lives. In short, it is a matter of citizen empowerment.

In healthcare this trend translates into “patient empowerment.” The report by the French Medical Council and the health report for the General Council for the Economy both advocate a new healthcare model in the form of a “health democracy” based on the “6P health”: Preventative, Predictive, Personalized, Participatory, Precision-based, Patient-focused.

“Humanistic AI” for the new treatment pathway

The following extract from the January 2018 CNOM report asserts that healthcare improvements provided by AI must focus on the patient, ethics and building a trust-based relationship between patient and caregiver: “Keeping secrets is the very basis of people’s trust in doctors. This ethical requirement must be incorporated into big data processing when building algorithms.” This is also expressed in the recommendations set forth in Cédric Villani’s report on Artificial Intelligence. Specifically, humanistic artificial intelligence for healthcare and society would include three components to ensure “responsible improvements”: responsibility, measurability and native ethics.

The complexity of medical devices and processes, the large number of parties involved, and the need to instantly access huge quantities of highly secure data require the use of AI as a Trusted Service platforms (AIaaTS), which natively integrate all the virtues of “sustainable digital technology.” AIaaTS is based on a data vault that incorporates the three key aspects of digital cognitive technologies, perception, reasoning and action, and is not limited to only deep or machine learning.

Guarantees of trust, ethics, measurability and responsibility would rely on characteristics that use tools of the new digital society. Native compliance with GDPR, coupled with strong user authentication systems would make it possible to ensure patient safety. The blockchain, with its smart contracts, can also play a role by allowing for enhanced notarization of administrative management and medical procedures. Indicators for ethics and explainability already exist to avoid the black-box effect as much as possible. Other indicators, similar to the Value Based Healthcare model developed by Harvard University Professor Michael Porter, measure the results of AI in healthcare and its added value for the patient.

All these open-ended questions allow us to reflect on the future of AI. The promised cognitive and functional improvements must be truly responsible improvements and must not be made at the expense of alienating individuals and society. If AI can be described as a machine for improvements, humanistic AI is a machine for responsible improvements.

 

RAMSES

RAMSES: Time keeper for embedded systems

I’MTech is dedicating a series of articles to success stories from research partnerships supported by the Télécom & Société Numérique Carnot Institute (TSN), to which Télécom ParisTech belongs.

[divider style=”normal” top=”20″ bottom=”20″]

Belles histoires, Bouton, CarnotEmbedded computing systems are sometimes responsible for performing “critical” functions. In the transport industry, they sometimes prevent collisions between two vehicles. To help design these important systems, Étienne Borde, a researcher at Télécom ParisTech specialized in embedded systems, developed RAMSES. This platform gives developers the tools they need to streamline the design process for these systems. Its potential for various applications in the industrial and transport sectors and robotics has been recognized by Télécom & Société Numérique Carnot Institute, which has made it a part of its technological platform.

 

What is the purpose of the RAMSES platform?

Étienne Borde: RAMSES is a platform that helps design critical real-time embedded systems. This technical term refers to embedded systems that have a significant time component: if a computer operation takes longer than planned, a critical system failure could occur. In terms of the software, time is managed by a real-time operating system. RAMSES automates the configuration of this system while ensuring the system’s time requirements are met.

What sectors could this type of system configuration support be used for?

EB: The transport sector is a prime candidate. We also have a case study for the railway sector that shows what the platform could contribute in this field. RAMSES is used to estimate the worst data transmission time for a train’s control system. The most critical messages transmitted ensure that the train does not collide with another train. For safety reasons, the calculations are carried out using three computing units at the back of the train and three computing units at the front of the train. What RAMSES offers is a better control of latency and better management of the flow of transmission operations.

How does RAMSES help improve the configuration of critical real-time embedded systems?

EB: RAMSES is a compiler of the AADL language. This language is used to describe computer architectures. The basic principle of AADL is to define categories of software or hardware components that correspond to physical objects used in the everyday life of computer scientists or electronic engineers. An example of one of these categories is that of processors: AADL can describe the computer’s calculation unit by its parameters and frequency.  RAMSES helps assemble these different categories to represent the system with different levels of abstraction. This explains how the platform got its name: Refinement of AADL Models for Synthesis of Embedded Systems.

How does a compiler like RAMSES benefit professionals?

EB: Professionals currently develop their systems manually using the programming language of their choice, or generate this code using a model. They can assess the data transmission time on the final product, but with poor traceability in relation to the initial model. If a command takes longer than expected, it is difficult for the developers to isolate the step causing the problem. RAMSES generates intermediate representations as it progresses, analyzing the time associated with each task to ensure no significant deviations occur. As soon as an accumulation of mechanisms present a major divergence in relation to the set time constraints, RAMSES alerts the professional. The platform can indicate which steps are causing the problem and help correct the AADL code.

Does this mean RAMSES is primarily a decision-support tool?

EB: Decision support is one part of what we do. Designing critical real-time embedded systems is a very complex task. Developers do not have all the information about the system’s behavior in advance. RAMSES does not eliminate all of the uncertainties, but it does reduce them. The tool makes it possible to reflect on the uncertainties to decide on possible solutions. The alternative to this type of tool is to make decisions without enough analysis. But RAMSES is not used for decision support only. The platform can also be used to improve systems’ resilience, for example.

How can optimizing the configuration impact the system’s resilience?

EB: Recent work by the community of researchers in systems security looks at mixed criticality. The goal is to use multi-core architectures to deploy critical and less critical functions on the same computing unit. If the time constraints for the critical functions are exceeded, the non-critical functions are degraded. The computing resources made available by this process are then used for critical functions, thereby ensuring their resilience.

Is this a subject you are working on?

EB: Our team has conducted work to ensure that critical tasks will always have enough resources, come what may. At the same time, we are working on the minimum availability of resources for less critical functions. This ensures that the non-critical functions are not degraded too often. For example, for a train, this ensures that the train does not stop unexpectedly due to a reduction in computational resources. In this type of context, RAMSES assesses the availability of the functions based on their degree of criticality, while ensuring enough resources are available for the most critical functions.

Which industrial sectors could benefit from the solutions RAMSES offers?

EB: The main area of application is the transport sector, which frequently uses critical real-time embedded systems. We have partnerships with Thales, Dassault and SAFRAN in avionics, Alstom for the railway sector and Renault for the automotive sector. The field of robotics could be another significant area of application. The systems in this sector have a critical aspect, especially in the context of large machines that could present a hazard to those nearby, should a failure occur. This sector could offer good use cases.

 

[divider style=”normal” top=”20″ bottom=”20″]

A guarantee of excellence in partnership-based research since 2006

The Télécom & Société Numérique Carnot Institute (TSN) has been partnering with companies since 2006 to research developments in digital innovations. With over 1,700 researchers and 50 technology platforms, it offers cutting-edge research aimed at meeting the complex technological challenges posed by digital, energy and industrial transitions currently underway in in the French manufacturing industry. It focuses on the following topics: industry of the future, connected objects and networks, sustainable cities, transport, health and safety.

The institute encompasses Télécom ParisTech, IMT Atlantique, Télécom SudParis, Institut Mines-Télécom Business School, Eurecom, Télécom Physique Strasbourg and Télécom Saint-Étienne, École Polytechnique (Lix and CMAP laboratories), Strate École de Design and Femto Engineering.

[divider style=”normal” top=”20″ bottom=”20″]

mobility

20 terms for understanding mobility

Increasingly linked to digital technology, mobility is becoming more complex and diversified. Autonomous driving, multimodality, rebound effect and agile method are all terms used in the study of new forms of mobility. The Fondation Mines-Télécom has put together a glossary of 20 terms to help readers understand the issues involved in this transformation and clarify the ideas and concepts explained in its 10th booklet.

 

Active mobility A form of transport that uses only the physical activity of the human being for locomotion. Cycling, walking, roller skating, skateboarding etc.

Agile development Use of agile methods to create projects based on iterative, incremental and adaptive development cycles.

API  Application programming interface. A set of methods and tools through which a software program provides services to other software programs.

Ergonomics Scientific study of the relationship between human beings and their work tools, methods and environment. Its aim is to develop systems that provide maximum comfort, safety and efficiency.

Explanation interviews Interviews aimed at establishing as detailed a description as possible about a past activity.

Flooding Availability of more resources than necessary.

Free floating Fleet of self-service vehicles available for use without a home station.

Gentrification Urban phenomenon whereby wealthier individuals appropriate a space initially occupied by less privileged residents.

Intermodality Use of several modes of transport during a single journey. Not to be confused with multimodality.

IoT Internet of Things

L4, L5 High levels of autonomous driving. L4: the driver provides a destination or instructions and may not be in the vehicle. L5: Fully autonomous driving in all circumstances, without help from the driver.

MaaS Mobility as a service. This concept was formalized by professionals at the 2014 ITS European Congress in Helsinki and through the launch of the MaaS Alliance.

Modality Used to describe a specific mode of transport characterized by the vehicles and infrastructures used.

Multimodality Presence of several modes of transport between different locations. Not to be confused with intermodality.

PDIE Inter-company travel plan (French acronym for Plan de Déplacement Inter-Entreprise). Helps make individual company travel plans (PDE) more effective by grouping together several companies and pooling their needs.

Rebound effect Economic term explaining the rise in consumption that occurs when the limits of using a technology are reduced.

Soft mobility Sometimes used as an exact synonym for non-motorized (and thereby active) forms of mobility. The term “soft” refers to environmental sustainability relating to eco-mobility: reducing noise, limiting pollution etc. It sometimes also includes motorized or assisted forms of transport based on technologies that do not rely on oil.

Solo car use Term used to describe a driver alone in his/her car.

Transfer An often expensive step during which merchandise or passengers are transferred from one vehicle to another. As short a transshipment as possible is desirable.

TCSP – Acronym used for exclusive right-of-way transit in France.

 

scripts

Computerizing scripts: losing character?

With the use of the Internet and new technologies, we rely primarily on writing to communicate. This change has led to a renewed interest in graphemics, the linguistic study of writing. Yannis Haralambous, a researcher in automatic language processing and text mining at IMT Atlantique helped organize the /gʁafematik/ conference held in June 2018. In this article, we offer a look at the research currently underway in the field of graphemics and address questions about the future of computerized scripts in Unicode encoding, that will decide the future of languages and their characters…

 

Graphemics. Never heard this word before? That’s not surprising. The word refers to a linguistic study similar to phonology, except that it focuses on a language’s writing system, an area that was long neglected as linguistic study tended to focus more on the spoken than on the written word. Yet the arrival of new technologies, based strongly on written text, has changed all that and explains the renewed interest in the study of writing. “Graphemics is of interest to linguists just as much as it is to computer engineers, psychologists, researchers in communication sciences and typographers,” explains Yannis Haralambous, a researcher in computerized language processing and text mining at IMT Atlantique. “I wanted to bring all these different fields together by organizing the/gʁafematik/ conference on June 14 and 15, 2018 on graphemics in the 21st century.”

Among those in attendance were typographers who came to talk about the readability of typographic characters on a screen. Based on psycholinguistic field studies, they work to design fonts with specific characters for children or people with reading difficulties. At the same time, several researchers presented work on the use of emojis. This research reveals how these little pictographs no larger than letters can enrich the text with their emotional weight or replace words and even entire sentences, changing the way we write on social networks. In addition, the way we type with keyboards and the speed of our typing could be used to create signatures for biometric identification, with possible applications in cyber-security or in the diagnosis of disorders like dyslexia.

Yet among all the research presented at this conference, one leading actor in written language stole the show: Unicode. Anyone who types on a computer, whether they know it or not, is constantly dealing with this generalized and universal character encoding. But its goal of integrating all the world’s languages is not without consequences on the development of writing systems and particularly for “sparse” modern languages and dead languages.

Questioning the encoding system’s dominance  

Type a key on your keyboard, for example the letter “e”. A message is sent to your machine in the form of a number, a numeric code, received by the keyboard driver and immediately converted into digital information called Unicode character. This is the information that is then stored or transmitted. At the same time, the screen driver interprets this Unicode character and displays an image of the letter “e” in your text processing area. “If you change the virtual keyboard in your machine’s system preferences to Russian or Chinese, the same key on the same physical keyboard will produce different characters from other writing systems proposed by Unicode,” explains Yannis Haralambous.

Unicode is the result of work by a consortium of actors in the IT industry like Microsoft, Apple and IBM. Today it is the only encoding system that has been integrated worldwide. “Because it is the only system of its kind, it bears all the responsibility for how scripts develop,” the researcher explains. By accepting to integrate certain characters into its directory while rejecting others, Unicode controls the computerization of scripts. While Unicode’s dominance does not pose any real threat to “closed” writing systems like the French language, in which the number of characters does not vary, “open” writing systems are likely to be affected. “Every year, new characters are invented in the Chinese language, for example, and the Unicode consortium must follow these changes,” the researcher explains. But the most fragile scripts are those of dead and ancient languages. Once the encoding proposals made by reports from enthusiasts or specialists have been accepted by the Unicode consortium, they become final. The recorded characters can no longer be modified or deleted. For Yannis Haralambous, this is incomprehensible. “Errors are building up and will forever remain unchanged!

Also, while the physical keyboard does not naturally lend itself to the computerized transcription of certain languages, like Japanese or Chinese, with thousands of different characters, the methods used for phonetic input and the system that suggests the associated characters help by-pass this initial difficulty. Yet certain Indian scripts and Arabic script are clearly affected by the very principle of cutting the script up into characters: “The same glyphs can be combined in different ways by forming ligatures that change the meaning of the text,” Yannis Haralambous explains. But Unicode does not share this opinion: because they are incompatible with its encoding system, these ligatures are seen as purely aesthetic and therefore non-essential… This results in a loss of meaning for some scripts.  But Unicode also plays an important role in a completely different area.

A move towards the standardization of scripts?

While we learn spoken language spontaneously, written language is taught. “In France, the national education system has a certain degree of power and contributes to institutionalizing the written language, as do dictionaries or the Académie Française,” says Yannis Haralambous.  In addition to language policies, Unicode strongly affects how the written language develops. “Unicode does not seek to modify or to regulate, but rather implement. However, due to its constraints and technical choices, the system has the same degree of power as the institutions.” The researcher lists the trio that, in his opinion, offer a complete view of writing systems and their development: the linguistic study of the script, its institutionalization, and Unicode.

The example of Malayalam writing, in the Indian state of Kerala, clearly shows the important role character encoding and software play in how a population adopts a writing system. In the 1980s, the government implemented a reform to simplify the Malayalam ligature script to make it easier to print and computerize. Yet the population kept its handwritten script with complex ligatures. When Unicode emerged in 1990, Mac and Windows operating systems only integrated the simplified institutional version. What changed things was the simultaneous development of open source software, which made the script with traditional ligatures available on computers. “Today, after a period of uncertainty at the beginning of the 2000s, the traditional writing has made a strong comeback,” the researcher observes, emphasizing the unique nature of this case: “The writing reforms imposed by a government are usually definitive, unless a drastic political change occurs.”

Unicode Malayalam graphématique, scripts

Traditional Malayalam script. Credit: Navaneeth Krishnan S

 

While open source software and traditional writing won in the case of Malayalam script, are we not still headed towards the standardization of writing systems due to the computerization of languages? “Some experts claim that in a century or two, we will speak a mixture of different languages with a high proportion of English, but also Chinese and Arabic. And the same for writing systems, with Latin script finally taking over… But in reality, we just don’t know,” Yannis Haralambous explains. The researcher also points out that the transition to Latin script has been proposed in Japan and China for centuries without success.  “The mental patterns of Chinese and Japanese speakers are built on the characters: In losing the structure of characters, they would lose the construction of meaning…” While language does contribute to how we perceive the world, the role of the writing system in this process has yet to be determined. This represents a complex research problem to solve. As Yannis Haralambous explains, “it is difficult to separate spoken language from written language to understand which influences our way of thinking the most…”

GDPR

Four flagship measurements of the GDPR for the economy

Patrick Waelbroeck, Professor in Economic Sciences, Télécom ParisTech, Institut Mines-Télécom (IMT)

[divider style=”normal” top=”20″ bottom=”20″]

[dropcap]N[/dropcap]umerous scandals concerning data theft or misuse that have shaken the media in recent years have highlighted the importance of data protection. The implementation of the General Data Protection Regulation (GDPR) in Europe is designed to do just that through four flagship measurements: seals of trust, accountability, data portability and pseudonymization. We explain how.

Data: potential negative externalities

The misuse of data can lead to identity theft, cyberbullying, disclosure of private information, fraudulent use of bank details or, less serious but nevertheless condemnable, price discrimination or the display of unwanted ads whether they are targeted or not.

All these practices can be assimilated to “negative externalities”. In economics, a negative externality results from a failing in the market. It occurs when the actions of an economic actor have a negative effect on other actors with no offset linked to a market mechanism.

In the case of the misuse of data, negative externalities are caused for the consumer by companies that collect too much data in relation to the social optimum. In terms of economics, their existence justifies the protection of personal data.

The black box society

The user is not always aware of the consequences of data processing carried out by companies working in digital technology, and the collection of personal data can generate risks if an internet user leaves unintentional records in their browser history.

Indeed, digital technology completely changes the conditions of exchange because of the information asymmetry that it engenders, in what Frank Pasquale, a legal expert at the University of Maryland, calls the “black box society”: users of digital tools are unaware of both the use of their personal data and the volume of data exchanged by the companies collecting it. There are three reasons for this.

Firstly, it is difficult for a consumer to check how their data is used by the companies collecting and processing it, and it is even more difficult to know whether this use is compliant with legislation or not. This is even more true in the age of “Big Data”, where separate databases containing little personal information can be easily combined to identify a person. Secondly, an individual struggles to technically assess the level of computer security protecting their data during transmission and storage. Thirdly, information asymmetries undermine the equity and reciprocity in the transaction and create a feeling of powerlessness among internet users, who find themselves alone facing the digital giants (what sociologists call informational capitalism).

Seals and marks of trust

The economic stakes are high. In his work that earned him the Nobel Prize in Economic Sciences, George Akerlof showed that situations of information asymmetry could lead to the disappearance of markets. The digital economy is not immune to this theory, and the deterioration of trust creates risks that may lead some users to “disconnect”.

In this context, the GDPR strongly encourages (but does not require) seals and marks of trust that allow Internet users to better comprehend the dangers of the transaction. Highly common in the agri-food industry, these seals are marks such as logos, color codes and letters that prove compliance with regulations in force. Can we count on these seals to have a positive economic impact for the digital economy?

What impacts do seals have?

Generally speaking, there are three positive economic effects of seals: a signaling effect, price effect and an effect on quantity.

In economics, seals and other marks of confidence are apprehended by signaling theory. This theory aims to solve the problems of information asymmetry by sending a costly signal that the consumer can interpret as proof of good practice. The costlier the signal, the greater its impact.

Economic studies on seals show that they generally have a positive impact on prices, which in turn boosts reputation and quantities sold. In the case of seals and marks of trust for personal data protection, we can expect a greater effect on volume than on prices, in particular for non-retail websites or those offering free services.

However, existing studies on seals also show that the economic impacts are all the more substantial when the perceived risks are greater. In the case of agri-food certifications, health risks can be a major concern; but risks related to data theft are still little known about among the general public. From this point of view, seals may have a mixed effect.

Other questions also remain unanswered: should seal certification be carried out by public or private organizations? How should the process of certification be financed? How will the different seals coexist? Is there not a risk of confusion for internet users? If the seal is too expensive, is there not a danger that small businesses won’t be able to become certified?

In reality, less than 15% of websites among the top 50 for online audience rates currently display a data protection seal. Will GDPR change this situation?

Accountability and fines

The revelations of the Snowden affair on the large-scale surveillance by States have also created a sense of distrust towards operators of the digital economy. To the extent that, in 2015, 21% of Internet users were prepared not to share any information at all, whereas this was true of only 5% in 2009. Is there a societal cost to the “everything for free” era on the Internet?

Instances of customer data theft among companies such as Yahoo, Equifax, eBay, Sony, LinkedIn or, more recently, Cambridge Analytica and Exactis, number in the millions. These incidents are too rarely followed up by sanctions. The GDPR establishes a principle of accountability which forces companies to be able to demonstrate that their data processing is compliant with the requirements laid down by the GDPR. The regulation also introduces a large fine that can be as much as 4% of worldwide consolidated turnover in the event of failure to comply.

These measures are a step in the right direction because people trust an economic system if they know that unacceptable behavior will be punished.

Data portability

The GDPR establishes another principle with an economic aim: data portability. Like the mobile telephony sector where number portability has helped boost competition, the regulation aims to generate similar beneficial effects for internet users. However, there is a major difference between the mobile telephony industry and the internet economy.

Economies of scale in data storage and use and the existence of network externalities on multi-sided online platforms (platforms that serve as an intermediary between several groups of economic players) have created internet monopolies. In 2017, for example, Google processed more than 88% of online searches in the world.

In addition, the online storage of information benefits users of digital services by allowing them to connect automatically and save their preferences and browser history. This creates a “data lock-in” situation resulting from the captivity of loyal users of the service and characterized by high costs of changing. Where can you go if you leave Facebook? This situation allows monopolizing companies to impose conditions of use of their services, facilitating large-scale exploitation of customer data, sometimes at the users’ expense. Consequently, the relationship between data portability and competition is like a “Catch-22”: portability is supposed to create competition, but portability is not possible without competition.

Pseudonymization and explicit consent

The question of trust in data use also concerns the economic value of anonymity. In a simplistic theory, we can suppose that there is an arbitration between economic value on the one hand and the protection of privacy and anonymity on the other. In this theory there are two extreme situations: one in which a person is fully identified and likely to receive targeted offers, and another in which the person is anonymous. In the first case, there is maximum economic value, while in the second the data is of no value.

If we move along the scale towards the targeting end, economic value is increased to the detriment of the protection of privacy. Conversely, if we protect privacy, the economic value of the data is reduced. This theory does not account for the challenges raised by trust in data use. To develop a long-term client relationship, it is important to think in terms of risks and externalities for the customer. There is therefore an economic value to the protection of privacy, which revolves around long-term relationships, the notion of trust, the guarantee of freedom of choice, independence and the absence of discrimination.

The principles of pseudonymization and explicit consent in the GDPR adopt this approach. Nevertheless, the main actors have to play along and comply, something which still seems far from being the case: barely a month after the date of effect of the regulations, the Consumer Council of Norway (Forbrukerrådet), an independent organization funded by the Norwegian Government, accused Facebook and Google of manipulating internet users to encourage them to share their personal information.

Patrick Waelbroeck, Professor in Economic Sciences, Télécom ParisTech, Institut Mines-Télécom (IMT)

The original version of this article (in French) was published on The Conversation.

Also read on I’MTech :

[one_half]

[/one_half]

[one_half_last]

[/one_half_last]

Empathic

AI to assist the elderly

Projets européens H2020Caring and expressive artificial intelligence? This concept that seems to come straight from a man-machine romance like the movie “Her”, is in fact at the heart of a Horizon 2020 project called EMPATHIC. The project aims to develop software for a virtual and customizable coach for assisting the elderly. To learn more, we interviewed the project’s Scientific Director for Télécom SudParis and expert in voice recognition, Dijana Petrovska-Delacretaz.

[divider style=”normal” top=”20″ bottom=”20″]

The original version of this article (in French) was published on the Télécom SudParis website.

[divider style=”normal” top=”20″ bottom=”20″]

What is the goal of the Empathic project?

Dijana Petrovska-Delacretaz: The project’s complete title is “Empathic Expressive Advanced Virtual Coach to Improve Independent Healthy Life-Years of the Elderly”. The goal is to develop a virtually advanced, empathetic and expressive coach to assist the elderly in daily life. This interface equipped with artificial intelligence and featuring several options would be adaptable, customizable and available on several types of media: PC, smartphone or tablet. We want to use a wide range of audiovisual information and interpret it to take the users’ situation into account and try to respond and interact with them in the most suitable way possible to offer them assistance. This requires us to combine several types of technology, such as signal processing and biometrics.

How will you design this AI?

DPD: We will combine signal processing and speech, face and shape recognition using “deep networks” (networks of deep artificial neurons). In other words, we will reproduce our brain’s structure using computer and processor calculations. This new paradigm has become possible thanks to the massive increase in storage and calculation capacities. They allow us to do much more in much less time.

We also use 3D modeling technology to develop avatars with more expressive faces–whether it be a human, cat, robot, hamster or even a dragon–that can be adapted to the user’s preferences to facilitate the dialogue between the user and virtual coach.   The most interesting solution from a biometric and artificial intelligence standpoint is to include as many options as possible: from using the voice and facial expressions to recognizing emotions.  All of these aspects will help the virtual coach recognize its user and have an appropriate dialogue with him or her.

Why not create a robot assistant?

The robot NAO is used at Télécom SudParis in work on recognition and gesture interpretation.

The robot NAO is used at Télécom SudParis in work on recognition and gesture interpretation.

DPD: The software can be fully integrated into a robot like Nao, which is already used here on the HadapTIC and EVIDENT platforms. This is certainly a feasible alternative: rather than have a virtual coach on a computer or smartphone, the coach can be there in person. The advantage with a virtual coach is that it is much easier to bring with me on my tablet when I travel.

In addition, from a security standpoint, a robot like Nao is not allowed access everywhere. We really want to develop a system that is very portable and not bulky. The challenge is to combine all of these types of technology and make them work together so we can interact with the artificial intelligence based on scenarios that are best suited to an elderly individual on different types of devices.

Who is participating in the Empathic project?

DPD: The Universidad del Pais Vasco in Bilbao is coordinating the project and is contributing to the dialogue aspect. We are complementing this aspect by providing the technology for vocal biometrics and emotional and face recognition, as well as the avatar’s physical appearance. The industrial partners involved in the project are Osatek, Tunstall, Intelligent Voice and Acapela. Osatek is the largest Spanish company specializing in connected objects for patient monitoring and home care. Tunstall also specializes in this type of material. Intelligence Voice and Acapela, on the other hand, will primarily provide assistance in voice recognition and speech synthesis. In addition, Osatek and Intelligent Voice will be responsible for ensuring the servers are secure and storing the data. The idea is to create a prototype that they will then be able to use.

Finally, the e-Seniors association, the Seconda Universidad of Italy and the Oslo University Hospital will provide the 300 users-testers to test the EMPATHIC interface throughout the project. This provides us with significant statistical validity.

What are the major steps in the project?

DPD: The project was launched in November 2017 and our research will last three years. We are currently preparing for an initial acquisition stage which take place in the spring of 2018. For this first stage, we will make recordings with a “Wizard of Oz” system in the EVIDENT living lab at Télécom SudParis. With this system, we can simulate the virtual agent with a person who is located in another room in order to collect data on the interaction between the user and the agent. This system allows us to create more complex scenarios that our artificial intelligence will then take into account. This first stage will also provide us with additional audiovisual data and allow us to select the best personifications of virtual coaches to offer to users. I believe this is important.

What are the potential opportunities and applications?

DPD: The research project is intended to produce a prototype. The idea behind having industrial partners is for them to quickly obtain and adapt this prototype to suit their needs. The goal of this project is not to produce a specific innovation, but rather to combine and adapt all these different types of technology so that they reach their potential through specific cases—primarily for assisting the elderly.

Another advantage of EMPATHIC is that it reveals many possible applications in other areas: including video games and social networks.  A lot of people interact with avatars in virtual worlds today–like in the video game Second Life, which is where one of our avatars, Bob the Hawaiian, comes from. EMPATHIC could therefore definitely be adapted outside the medical sector.

Do not confuse virtual and reality

The “Uncanny Valley” is a central concept in artificial intelligence and robotics. It theorizes that if a robot or AI possess a human form (whether physical or virtual), it should not resemble humans too closely–unless this resemblance is perfectly achieved from every angle.  “As soon as there is a strong resemblance with humans, the slightest fault is strange and disturbing. In the end, it becomes a barrier in the interactions with the machine,” explains Patrick Horain, a research engineer at Télécom SudParis, specialized in digital imaging.

Dijana Petrovska-Delacretaz has integrated this aspect and is developing the EMPATHIC coaches accordingly: “It is important to keep this in mind. In general, it is a scientific challenge to make a photo-realistic face that actually looks like a loved one without being able to distinguish the real from the fake. But in the context of our project, it is a problem. It could be disturbing or confusing for an elderly person to interact with an avatar that looks like someone real or even a loved one. It is always preferable to propose virtual coaches that are clearly not human.”

[divider style=”normal” top=”20″ bottom=”20″]

Do not confuse virtual and reality

The “Uncanny Valley” is a central concept in artificial intelligence and robotics. It theorizes that if a robot or AI possess a human form (whether physical or virtual), it should not resemble humans too closely–unless this resemblance is perfectly achieved from every angle.  “As soon as there is a strong resemblance with humans, the slightest fault is strange and disturbing. In the end, it becomes a barrier in the interactions with the machine,” explains Patrick Horain, a research engineer at Télécom SudParis, specialized in digital imaging.

Dijana Petrovska-Delacretaz has integrated this aspect and is developing the EMPATHIC coaches accordingly: “It is important to keep this in mind. In general, it is a scientific challenge to make a photo-realistic face that actually looks like a loved one without being able to distinguish the real from the fake. But in the context of our project, it is a problem. It could be disturbing or confusing for an elderly person to interact with an avatar that looks like someone real or even a loved one. It is always preferable to propose virtual coaches that are clearly not human.”

[divider style=”normal” top=”20″ bottom=”20″]

 

Ttool

TTool offers security, safety and performance for embedded systems

I’MTech is dedicating a series of articles to success stories from research partnerships supported by the Télécom & Société Numérique Carnot Institute (TSN), to which Télécom ParisTech belongs.

[divider style=”normal” top=”20″ bottom=”20″]

Belles histoires, Bouton, CarnotBy combining three areas for assisting the design of embedded systems, the TTool platform is seen as a safety net for industry. The platform, which originated in the laboratories of Télécom ParisTech, has now become one of the technology platforms Carnot TSN offers to companies. TTool helps them design embedded systems that are more secure, protected from cyberattacks and more efficient. Ludovic Apvrille, the Telecom ParisTech researcher behind this platform, explains the tool’s strengths and how it works.

 

What does the TTool platform you developed do?

Ludovic Apvrille: It is a design tool for embedded systems for creating models and running tests. On the one hand, it takes into account operation safety, in other words, it ensures the systems do not cause any risks to humans or have serious economic impacts. On the other hand, it offers a cybersecurity aspect by preventing certain attacks. This platform’s strength is that it offers both these aspects, whereas industrial tools only offer one or the other.

How does TTool ensure a system operates safely?

LA: Embedded system designers want to know whether a specific error could occur. These specific errors can be avoided by voluntarily degrading the system’s operation level when there is a high likelihood of the error occurring. This makes the system less efficient, but at least it does not endanger the user. It allows safety features to be integrated directly into the system itself. TTool helps implement these features. The developers create models of their system, enter the aspects they want to verify, and then just press a button. TTool does the rest: the platform analyzes the system and the potential error, then reports whether safety can be ensured.

Is this different from the cybersecurity aspect?

LA: Yes, the cybersecurity part is different. Embedded systems are vulnerable to different types of attacks, especially those that are used to read or inject data into the system’s communication links. We use TTool to study the system’s vulnerabilities to this type of attack: TTool can detect this type of attack and then automatically add software and hardware components to counter them. These components can ensure the data’s confidentiality, prevent a hacker from posing as a user, or indicate whether data has been altered by a hacker.

In specific terms, how can manufacturers use a platform like TTool?

LA: A good example is the joint Nokia Bell Labs and Télécom ParisTech laboratory inaugurated this year on June 25. TTool is one of the tools Nokia uses to respond to problems. In this case—due to the platform’s ability to quickly analyze a system’s model—it provides the company with information on the expected performance. We are specifically interested in the latency related to processing data. Nokia works with 5G encoders/decoders, components which encode and decode data packets sent by fifth generation mobile technology. TTool indicates how long it takes for a data sample to be encoded or decoded based on the choice of architecture.

How does TTool work?

LA: TTool is based on three modeling environments. The first is Diplodocus, which partitions the functions into software and hardware features. In the Nokia example, signal processing is carried out in part by the base stations. If there are major changes in the processing functions, the base stations can no longer provide this feature. The equipment providers therefore want to move some of the processing to the cloud. Since this is a costly operation, they want to minimize the number of calculations. In this case, the functions performed on the hardware and by the software must be optimized. Next, TTool is based on the AVATAR environment, which focuses on creating the embedded software: modeling, testing, then code generation. Finally, there is the SySML-Sec environment, which helps the developer add security functions to the embedded system for the entire development cycle.

Which sectors benefit most from this platform?

LA: It’s fairly diverse. TTool is currently being used for the European project H2020 AQUAS for two case studies on critical systems. In these cases, a security study must be carried out on both the safety of the operations and the performance. One is with Siemens, to add security to industrial motors. The other is for railway systems for automated doors—like the metro on line 1 in Paris. The objective is to detect if cyberattacks could affect the time the doors open or close. TTool can therefore be adapted to a wide range of embedded systems.

What are TTool’s main advantages?

LA: TTool offers a safety net during the development phase for embedded systems and it helps in the path to certification. It is therefore a platform that can increase trust in a product, beginning at the design phase. This does not mean the embedded systems can avoid the testing phase they are required to undergo, but there are less chances of finding errors. It should also be noted that TTool is free and open source software. This means that everyone can use the tools and modify the code to adapt it to their needs. The research team behind TTool intervenes when there is an issue they can work on to help companies improve their capacities.

 

[divider style=”normal” top=”20″ bottom=”20″]

A guarantee of excellence in partnership-based research since 2006

The Télécom & Société Numérique Carnot Institute (TSN) has been partnering with companies since 2006 to research developments in digital innovations. With over 1,700 researchers and 50 technology platforms, it offers cutting-edge research aimed at meeting the complex technological challenges posed by digital, energy and industrial transitions currently underway in in the French manufacturing industry. It focuses on the following topics: industry of the future, connected objects and networks, sustainable cities, transport, health and safety.

The institute encompasses Télécom ParisTech, IMT Atlantique, Télécom SudParis, Institut Mines-Télécom Business School, Eurecom, Télécom Physique Strasbourg and Télécom Saint-Étienne, École Polytechnique (Lix and CMAP laboratories), Strate École de Design and Femto Engineering.

[divider style=”normal” top=”20″ bottom=”20″]