Invenis

Invenis: machine learning for non-expert data users

Invenis went into incubation at Station F at the beginning of July, and has since been developing at full throttle. This start-up has managed to make a name for itself in the highly competitive sector of decision support solutions using data analysis. Its strength? Providing easy-to-use software aimed at non-expert users, which processes data using efficient machine learning algorithms.

 

In 2015, Pascal Chevrot and Benjamin Quétier, both in the Ministry of Defense at the time, made an observation that made them want to launch a business. They considered that the majority of businesses were using outdated digital decision support tools that were increasingly ill-suited to their needs. “On the one hand, traditional software was struggling to adapt to big data processing and artificial intelligence”, Pascal Chevrot explains. “On the other hand, there were expert tools that existed but were inaccessible to anyone that didn’t have significant technical knowledge.” Faced with this situation, the two colleagues founded Invenis in November 2015 and joined the ParisTech Entrepreneurs incubator. On July 3, 2017, less than two years later, they joined Station F, one of the biggest start-up campuses in the world located in the 13th arrondissement of Paris.

The start-up is certainly appealing: it aims to rectify the lack of available decision support tools with SaaS software (Software as a Service). Its goal is to make the value provided by data available to people that manipulate them every day in order to obtain information, but who are by no means experts. Invenis therefore targets professionals that know how to extract data and use it to obtain information, but who find themselves limited by the capabilities of the tools that they use when they want to go further. Through their solution, Invenis allows these professionals to carry out data processing using machine learning algorithms, simply.

Pascal Chevrot illustrates how simple it is to use, with an example. He takes two data sets and uploads them to Invenis: one is the number of sports facilities per activity and per department, and the other is the population by city in France. The user can then choose what kind of data processing they wish to perform from a library of modules. For example, they could first decide to group the different kinds of sports facilities (football stadiums, boules pitches, swimming pools, etc.) according to regions in France. In parallel, the software will then aggregate the number of inhabitants per commune in order to provide a population value on a regional scale. Once each of these actions has been completed, the user can then carry out an automated segmentation, or “clustering”, in order to classify regions into different groups according to the density of sports facilities in that particular region. In a few clicks, Invenis thus allows users to visualize the regions that have the highest number of sports facilities and those with a low number in relation to the population size, and which should therefore be invested in. Each process carried out on the data is done simply by dragging a processing module into the interface associated with the desired procedure and using this to create a full data processing session.

The user-friendly nature of the Invenis software lies in how simple it is to use these processing modules. Every action has been designed to be simple for the user to understand. The algorithms come from open source libraries Hadoop and Spark, which are references in the sector. “We then add our own algorithms to these existing algorithms, making them easier to manage”, highlights Pascal Chevrot.

For example, the clustering algorithm they use ordinarily requires a certain number of factors to be defined. Invenis’ processing module automatically calculates these factors using its proprietary algorithms. It does, however, allow expert users to modify these if necessary.

In addition to how simple it is to use, the Invenis program has other advantages, namely a close management of data access rights. “Few tools do this”, affirms Pascal Chevrot, before demonstrating the advantages of this function: “For some businesses, such as telecommunication operators, it’s important because they have to report to the CNIL (National Commission for Data Protection and Liberties) for the confidentiality of their data, and soon this will also be the case in Europe, with the arrival of GDPR. Not forgetting that more established businesses have implemented data governance over these questions.”

 

Revealing the value of data

Another advantage of Invenis is that it offers different frameworks. The start-up offers free trial periods to any data users who are using tools they are not satisfied with, along with the opportunity to talk to the technical management team who can demonstrate the tool’s capabilities and even develop proof of concept. However, the start-up also has a support and advice service for businesses that have identified a problem that they would like to solve using their data. “We offer clients guaranteed results, assisting them to resolve their problem with the intention of ultimately making them independent”, explains the co-founder.

It was within this second format that Invenis realized its most iconic proof of concept with CityTaps, another start-up from ParisTech Entrepreneurs that offers prepaid water meters. Using the Invenis software allowed CityTaps to look at three questions. Firstly, how do users consume water in terms of days of the week, size of household, season, etc.? Secondly, what is the optimal moment to warn a user that they need to top up their meter, and would they be quick to do this after receiving an alert SMS? And finally, how can we best predict temperature changes in the meters due to the weather? Invenis provided many responses to these questions by using their processing solutions on CityTaps’ data.

The case of CityTaps shows to what extent data management tools are crucial for companies. Machine learning and intelligent data processing are essential in generating value. However, these technologies can sometimes be difficult to access due to insufficient technical knowledge. Enabling businesses to access this value by reducing access costs in terms of skills is Invenis’ number one aim. As Pascal Chevrot concludes, the key is to provide “”.

algorithms

Ethics, an overlooked aspect of algorithms?

We now encounter algorithms at every moment of the day. But this exposure can be dangerous. It has been proven to influence our political opinions, moods and choices. Far from being neutral, algorithms carry their developers’ value judgments, which are imposed on us without our noticing most of the time. It is now necessary to raise questions about the ethical aspects of algorithms and find solutions for the biases they impose on their users.

 

[dropcap]W[/dropcap]hat exactly does Facebook do? Or Twitter? More generally, what do social media sites do? The overly-simplified but accurate answer is: they select the information which will be displayed on your wall in order to make you spend as much time as possible on the site. Behind this time-consuming “news feed” hides a selection of content, advertising or otherwise, optimized for each user through a great reliance on algorithms. Social networks use these algorithms to determine what will interest you the most. Without questioning the usefulness of these sites — this is most likely how you were directed to this article — the way in which they function does raise some serious ethical questions. To start with, are all users aware of algorithms’ influence on their perception of current events and on their opinions? And to take a step further, what impacts do algorithms have on our lives and decisions?

For Christine Balagué, a researcher at Télécom École de Management and member of CERNA (see text box at the end of the article), “personal data capturing is a well-known topic, but there is less awareness about the processing of this data by algorithms.” Although users are now more careful about what they share on social media, they have not necessarily considered how the service they use actually works. And this lack of awareness is not limited to Facebook or Twitter. Algorithms now permeate our lives and are used in all of the mobile applications and web services we use. All day long, from morning to night, we are confronted with choices, suggestions and information processed by algorithms: Netflix, Citymapper, Waze, Google, Uber, TripAdvisor, AirBnb, etc.

Are your trips determined by Citymapper? Or by Waze? Our mobility is increasingly dependent on algorithms. Illustration: Diane Rottner for IMTech

 

They control our lives,” says Christine Balagué. “A growing number of articles published by researchers in various fields have underscored the power algorithms have over individuals.” In 2015, Robert Epstein, a researcher at the American Institute for Behavioral Research, demonstrated how a search engine could influence election results. His study, carried out with over 4,000 individuals, demonstrated that candidates’ rankings in search results influenced at least 20 % of undecided voters. In another striking example, a study carried out by Facebook in 2012 on 700,000 of its users showed that people who had previously been exposed to negative posts posted predominantly negative content. Meanwhile, those who had previously been exposed to positive posts posted essentially positive content. This proves that algorithms are likely to manipulate individuals’ emotions without their realizing or being informed of it. What role do our personal preferences play in a system of algorithms of which we are not even aware?

 

The opaque side of algorithms

One of the main ethical problems with algorithms stems from this lack of transparency. Two users who carry out the same query on a search engine such as Google will not have the same results. The explanation provided by the service is that responses are personalized to best meet the needs of each of these individuals. But the mechanisms for selecting results are opaque. Among the parameters taken into account to determine which sites will be displayed on the page, over a hundred have to do with the user performing the query. Under the guise of trade secret, the exact nature of these personal parameters and how they are taken into account by Google’s algorithms is unknown. It is therefore difficult to know how the company categorizes us, determines our areas of interest and predicts our behavior. And once this categorization has been carried out, is it even possible to escape it? How can we maintain control over the perception that the algorithm has created about us?

This lack of transparency prevents us from understanding possible biases which could result from data processing. Nevertheless, these biases do exist and protecting ourselves from them is a major issue for society. A study by Grazia Cecere, an economist at Télécom École de Management, provides an example of how individuals are not treated equally by algorithms. Her work has highlighted discrimination between men and women in a major social network’s algorithms for associating interests. “In creating an ad for STEMs (sciences, technology, education, mathematics), we noticed that the software demonstrated a preference for distributing it to men, even though women show more interest for this subject,” explains Grazia Cecere. Far from the myth of malicious artificial intelligence, this sort of bias is rooted in human actions. We must not forget that behind each line of code, there is a developer.

Algorithms are used first and foremost to propose services, which are most often commercial in nature. They are thus part of a company’s strategy and reflect this strategy in order to respond to its economic demands. “Data scientists working on a project seek to optimize their algorithms without necessarily thinking about the ethical issues involved in the choices made by these programs,” points out Christine Balagué. In addition, humans have perceptions about the society to which they belong and integrate these perceptions, either consciously or unconsciously, in the software they develop. Indeed, value judgements present in algorithms quite often reflect the value judgments of their creators. In the example of Grazia Cecere’s work, this provides a simple explanation for the bias discovered, “An algorithm learns what it is asked to learn and replicates stereotypes if they are not removed.”

algorithms

What biases are hiding in the digital tools we use every day? What value judgments passed down from algorithm developers do we encounter on a daily basis? Illustration: Diane Rottner for IMTech.

 

A perfect example of this phenomenon involves medical imaging. An algorithm used to classify a cell as sick or healthy must be configured to make a comparative assessment of the number of false positives and false negatives. Developers must therefore decide to what extent it is tolerable for healthy individuals to receive positive tests in order to prevent sick individuals from receiving negative tests. For doctors, it is preferable to have false positives rather than false negatives while scientists who develop algorithms prefer false negatives to false positives, as scientific knowledge is cumulative. Depending on their own values, developers will privilege one of these professions.

 

Transparency? Of course, but that’s not all!

One proposal for combating these biases is to make algorithms more transparent. Since October 2016, the law for a digital republic, proposed by Axelle Lemaire, the former Secretary of State for Digital Affairs, requires transparency for all public algorithms. This law was responsible for making the higher education admission website (APB) code available to the public. Companies are also increasing their efforts for transparency. As of May 17, 2017, Twitter has allowed its users to see the areas of interest the site associates with them. But despite these good intentions, the level of transparency is far from sufficient for ensuring the ethical dimension. First of all, code understandability is often overlooked: algorithms are sometimes delivered in formats which make them difficult to read and understand, even for professionals. Furthermore, transparency can be artificial. In the case of Twitter, “no information is provided about how user interests are attributed,” observes Christine Balagué.

[Interests from Twitter
These are some of the interests matched to you based on your profile and activity.
You can adjust them if something doesn’t look right.]

Which of this user’s posts led to his being classified under “Action and Adventure,” a very broad category? How are “Scientific news” and “Business and finance” weighed in order to display content in the user’s Twitter feed?

 

To take a step further, the degree to which algorithms are transparent must be assessed. This is the aim of the TransAlgo project, another initiative launched by Axelle Lemaire and run by Inria. “It’s a platform for measuring transparency by looking at what data is used, what data is produced and how open the code is,” explains Christine Balagué, a member of TransAlgo’s scientific council. The platform is the first of its kind in Europe, making France a leading nation in transparency issues. Similarly, DataIA, a convergence institute for data science established on Plateau de Saclay for a period of ten years, is a one-of-a-kind interdisciplinary project involving research on algorithms in artificial intelligence, their transparency and ethical issues.

The project brings together multidisciplinary scientific teams in order to study the mechanisms used to develop algorithms. The humanities can contribute significantly to the analysis of the values and decisions hiding behind the development of codes. “It is now increasingly necessary to deconstruct the methods used to create algorithms, carry out reverse engineering, measure the potential biases and discriminations and make them more transparent,” explains Christine Balagué. “On a broader level, ethnographic research must be carried out on the developers by delving deeper into their intentions and studying the socio-technological aspects of developing algorithms.” As our lives increasingly revolve around digital services, it is crucial to identify the risks they pose for users.

Further reading Artificial Intelligence: the complex question of ethics

[box type=”info” align=”” class=”” width=””]

A public commission dedicated to digital ethics

Since 2009, the Allistene association (Alliance of digital sciences and technologies) has brought together France’s leading players in digital technology research and innovation. In 2012, this alliance decided to create a commission to study ethics in digital sciences and technologies: CERNA. On the basis of multidisciplinary studies combining expertise and contributions from all digital players, both nationally and worldwide, CERNA raises questions about the ethical aspects of digital technology. In studying such wide-ranging topics as the environment, healthcare, robotics and nanotechnologies, it strives to increase technology developers’ awareness and understanding of ethical issues.[/box]

 

 

 

human resources

How is technology changing the management of human resources in companies?

In business, digital technology is revolutionizing more than just production and design. The human resources sector is also being greatly affected; whether that be through better talent spotting, optimized recruitment processes or getting employees more involved with the company’s vision. This is illustrated though two start-ups incubated at ParisTech Entrepreneurs, KinTribe and Brainlinks.

 

Recruiters in large companies can sometimes store tens of thousands of profiles in their databases. However, it often difficult to make use of such a substantial pool of information using conventional methods. “It’s impossible to keep such a large file up-to-date, so the data often become obsolete very quickly”, states Chloé Desault, a former in-company recruiter and co-founder of the start-up KinTribe. “Along with Louis Laroche, my co-founder who was also formerly a recruiter, we aim to facilitate the use of these talent pools and improve the daily lives of recruiters”, she adds. The software solution enables recruitment professionals to create a recruitment pool using professional social networks. With KinTribe, they are able to create a usable database, in which they can perform complex searches in order to find the best person to contact for their need, from tens of thousands of available profiles. “This means they no longer have to waste time on people that do not correspond to the target in question”, affirms the co-founder.

The software’s algorithms can then process the collected data to produce a rating for the relevant market. This rating indicates to what extent a person is susceptible to an external recruitment offer. “70% of people on LinkedIn aren’t actively looking for a job, but would still consider a job offer if it was presented to them”, Louis Laroche explains. In order to identify these people, and to what extent they are likely to be interested, the algorithm is based on key values identified by recruiters. Age, field of work and duration of last employment are all factors that can influence how open someone is to a proposition.

One of the start-up’s next goals is to add new sources of data into the mix, allowing their users to go on other networks to find new talents. Multiplying the available data will also allow them to improve the market rating algorithms. “We want to provide recruiters with the best possible knowledge by aggregating the maximum amount of social data that we can”; summarizes the KinTribe co-founder.

Finally, the two entrepreneurs are also interested in other topics within in the field of recruitment. “As a start-up, we have to try to stay ahead of the curve and understand what the market will do next. We dedicated part of our summer to exploring the potential of a new co-optation product” Chloé concludes.

 

From recruitment to employee involvement

In human resources, software tools represent more than just an opportunity for recruiting new talent. One of their aims is also to get employees involved in the company’s vision, and to listen to them in order to pinpoint their expectations. The start-up Brainlinks was created for this very reason. Today, it offers a mobile app called Toguna for businesses with over 150 people.

The concept is simple: with Toguna, general management or human resources departments can ask employees a question such as: “What is your view of the company of the future?” or “What new things would you like to see in the office?” The employees, who remain anonymous on the app, can then select the questions they are interested in and offer responses that will be made public. If a response made by a colleague is interesting, other employees can vote for it, thus creating a collective form of involvement based on questions about life at work.

In order to make Toguna appeal to the maximum number of people, Brainlinks has opted for a smart, professional design: “Contributions are formatted by the person writing them; they can add an image and choose the font, etc.”, explains Marc-Antoine Garrigue, the start-up’s co-founder. “There is an element of fun that allows each person to make their contributions their own”, he continues. According to Marc-Antoine Garrigue, this feature has helped them reach an average employee participation rate of 85%.

Once the votes have been cast and the propositions collected, managers can analyze the responses. When a response is chosen, it is highlighted on the app, providing transparency in the inclusion of employee ideas. A field of development for the app is to continue improving the dialogue between managers and employees. “We hope to go even further in constructing a collective history: employees make contributions on a daily basis and in return, the management can explain the decisions they have made after having consulted these”, outlines the co-founder. This is an opportunity that could really help businesses to see digital transformation as a vehicle for creativity and collective involvement.

 

Fine particles are dangerous, and not just during pollution peaks

Véronique Riffault, IMT Lille Douai – Institut Mines-Télécom and François Mathé, IMT Lille Douai – Institut Mines-Télécom

[divider style=”normal” top=”20″ bottom=”20″]

[dropcap]T[/dropcap]he French Agency for Food, Environmental and Occupational Health and Safety (ANSES) released a new notice concerning air pollution yesterday. After having been questioned on the potential changes to norms for ambient air quality, particularly concerning fine particles (PM10 and PM2.5), the organization has highlighted the importance of pursuing work on implementing long-term public policies that promote the improvement of air quality. They recommend lowering the annual threshold value for PM2.5 to equal the recommendations made by the WHO, and introducing a daily threshold value for this pollutant. As the following data visualization shows, the problem extends throughout Europe.

Average concentrations of particulate matters whose aerodynamic diameter is below 2.5 micrometers (called “PM2.5” which makes up “fine particles” along with PM10) for the year 2012. Amounts calculated using measures from fixed air quality monitoring stations, shown in micrograms per m3 of air. Data source: AirBase.

The level reached in peak periods is indicated by hovering the mouse over a given circle, the sizes of which will vary depending on the amount. The annual average is also provided, detailing long-term exposure and the subsequent proven impact on health (particularly on the respiratory and cardio-vascular systems). It should be noted that the annual target value for the particles (PM2.5), as specified by European legislation is currently 25 µg/m3. The level will drop to 20 µg/m3 in 2020, whilst the WHO currently recommends an annual threshold of 10 µg/m3.

The data shown on this map correspond exclusively to a so-called “fundamental” site typology, examining not only urban environments but also rural ones, which are far from being influenced by nearby pollution (coming from traffic or industry). Airbase also collects data supplied by member states that use measuring methods that can vary depending on the site but always respect the data quality objectives and are specific to the pollutant (90% of data on PM2.5 is approved annually, with an uncertainty of ± 25%). This perhaps explains why certain regions show little or no data (Belarus, Ukraine, Bosnia-Herzegovina and Greece), keeping in mind that a single station cannot be representative of the air quality across an entire country (as is the case in Macedonia).

The PM2.5 shown here may be emitted directly into the atmosphere (these are primary particles) or formed by chemical reactions between gaseous pollutants in the atmosphere (secondary particles). The secondary formation of PM2.5 often stems from peaks in pollution at certain points in the year when the sources of the pollutants are most significant and in meteorological conditions which allow them to accumulate. Sources connected to human activity are mainly linked to combustion processes (such as engines in vehicles or the burning of biomass and coal for residential heating systems) and agricultural activity.

The above map shows that the threshold suggested by the WHO has been surpassed in a large majority of stations, particularly in Central Europe (Slovakia, South Poland) due to central heating methods, or in Northern Italy (the Po Valley), which has been affected by poor topographical and meteorological conditions.

Currently, only 1.16% of stations are recording measurements that are still within the WHO recommendations for PM2.5 (shown in light green on the map). On top of this, 13.6% of stations have already reached the future European limits to be set in 2020 (shown in green and orange circles).

This illustrates that a large section of the European population is being exposed to concentrations of particles that are harmful to health and that some significant efforts are to be made. In addition, when considering that the mass concentration of particulates is a good indicator of air quality, their chemical composition should not be forgotten. This is something which proves to be a challenge for health specialists and policymakers, especially in real time.

Véronique Riffault, Professor in Atmospheric Sciences, IMT Lille Douai – Institut Mines-Télécom and François Mathé, Professor-Researcher, President of the AFNOR X43D Normalization Commission “Ambient Atmospheres”, Head of Studies at LCSQA (Laboratoire Central de Surveillance de la Qualité de l’Air), IMT Lille Douai – Institut Mines-Télécom

 

The original version  of this article was published in French in The Conversation France.

 

 

auonomous cars, Guillaume Duc, chair C3S, connected cars

No autonomous cars without cybersecurity

Protecting cars from cyber-attacks is an increasingly important concern in developing smart vehicles. As these vehicles become more complex, the number of potential hacks and constraints on protection algorithms is growing. Following the example of the “Connected cars and cybersecurity” chair launched by Télécom ParisTech on October 5, research is being carried out to address this problem. Scientists intend to take on these challenges, which are crucial to the development of autonomous cars.

 

Connected cars already exist. From smartphones connected to the dashboard, to computer-aided maintenance operations, cars are packed with communicating embedded systems. And yet, they still seem to be a long way from the futuristic vehicles we’ve been dreaming up in our imagination. They do not (yet) all communicate with one another or with road infrastructures to provide warnings about dangerous situations for example. Cars are struggling to make the leap from “connected” to “intelligent”. And without intelligence, they will never become autonomous. Guillaume Duc, a research professor in electronics at Télécom ParisTech who specializes in embedded systems, perfectly sums up one of the hurdles to this development, “Autonomous cars will not exist until we are able to guarantee that cyber-attacks will not put a smart vehicle, its passengers or its environment in danger.”

Cybersecurity for connected cars is indeed crucial to their development. Whether rightly or wrongly, no authority will authorize the sale of increasingly intelligent vehicles without first guaranteeing that they will not be out of control on the roads. The topic is of such importance in the industry that researchers and manufacturers have teamed up to find solutions. A “Connected Cars and Cybersecurity” chair bringing together Télécom ParisTech, Fondation Mines-Télécom, Renault, Thalès, Nokia, Valéo and Wavestone was launched on October 5. According to Guillaume Duc, the specific features of connected cars make this is a unique research topic.

The security objectives are obviously the same as in many other systems,” he says, pointing to the problems of information confidentiality or certifying that information has really been sent by one sensor instead of another. “But cars have a growing number of components, sensors, actuators and communication interfaces, making them easier to hack,” he goes on to say. The more devices there are in a car, the more communication points it has with the outside world. And it is precisely these entry points which are the most vulnerable. However, these are not necessarily the instruments that first come to mind, like radio terminals or 4G.

Certain tire pressure sensors use wireless communication to display a possible flat tire on the dashboard. But wireless communication means that without an authentication system to ensure that the received information has truly been sent by this sensor, anyone can pretend to be this sensor from outside the car. And if you think sending incorrect information about tire pressure seems insignificant, think again. “If the central computer expects a value of between 0 and 10 from the sensor and you send it a negative number, for example, you have no idea how it will react,” explains the researcher. This could crash the computer, potentially leading to more serious problems for the car’s controls.

 

Adapting cybersecurity mechanisms for cars

The stakes are high for research on how to protect each of these communicating elements. These components have only limited computing power while algorithms to protect against attacks usually require high computing power. “One of the chair’s aims is to successfully adapt algorithms to guarantee security while requiring less computer resources,” says Guillaume Duc. This challenge goes hand in hand with another one, limiting latency in the car’s execution of critical decisions. Adding algorithms to embedded systems leads to increased computing time when an action is transmitted. But cars cannot afford to take longer to brake. Researchers therefore have their work cut out for them.

In order to address these challenges, they are looking to the avionics sector, which has been facing problems associated with the proliferation of sensors for years. But unlike planes, fleets of cars are not operated in an ultra-controlled environment. And in contrast to aircraft pilots, drivers are masters of their own cars and may handle them as they like. Cars are also serviced less regularly. It is therefore crucial to guarantee that cybersecurity tools installed in cars cannot be altered by their owners’ tinkering.

And since absolute security does not exist and “algorithms may eventually be broken, whether due to unforeseen vulnerabilities or improved attack techniques,” as the researcher explains, these algorithms must also be agile, meaning that they can be adapted, upgraded and improved without automakers having to recall an entire series of cars.

 

Resilience when faced with an attack

But if absolute security does not exist, where does this leave the 100% security guarantee against attacks, which is the critical factor in developing autonomous cars? In reality, researchers do not seek to protect against all possible attacks on connected cars. Their goal rather to ensure that even if an attack is successful, it will not prevent the driver or the car itself from remaining safe. And of course, this must be possible without having to brake suddenly on the highway.

To reach these objectives, researchers are using their expertise to develop resilience in embedded systems. The problem recalls that of critical infrastructures, such as nuclear power plants, which cannot simply shut down when under attack. In the case of cars, a malicious intrusion in the system must first be detected when it occurs. To do so, the vehicle’s behavior is constantly compared to previously-recorded behaviors which are considered normal. If an action is suspicious, it is identified as such. In the event of a real attack, it is crucial to guarantee that the car’s main functions (steering, brakes etc.)  will be maintained and isolated from the rest of the system.

Ensuring a car’s resilience from its design phase, resilience by design, is also the most important condition for cars to continually become more autonomous. Automakers can provide valuable insight for researchers in this area, by contributing to discussions about a solution’s technical acceptability or weighing in on economic issues. While it is clear that autonomous cars cannot exist without security, it is equally clear that they will not be rolled out if the cost of security makes them too expensive to find a market.

[box type=”info” align=”” class=”” width=””]

Cars: personal data on wheels!

The smarter the car, the more personal data it contains. Determining how to protect this data will also be a major research area for the “Connected cars and cybersecurity” chair. To address this question it will work in collaboration with another Télécom ParisTech chair, dedicated to “Values and policies of personal information” which also brings together Télecom SudParis and Télécom École de Management. This collaboration will make it possible to explore the legal and social aspects of exchanging personal data between connected cars and their environment. [/box]

ledgys, blockchain, vivatechnology

With Ledgys, blockchain becomes a tangible reality

Winner of the blockchain BNP jury prize at VivaTechnology last year, Ledgys hopes to reaffirm its relevance once more this year. From June 15 to 17, 2017, the company will be attending the event again to present its Ownest application. The application is aimed at both businesses and consumers, offering a simple solution for using blockchain technology in practical cases such as logistics management or an authenticity certificate for a handbag.

 

Beyond the fantasy of the blockchain, what is the reality of this technology? Ledgys, a start-up founded in April 2016 and currently incubated at Télécom ParisTech, answers this question in a very appealing way. By developing their app, Ownest, they offer professionals and individuals easy access to their blockchain portfolio. Users can therefore easily visualize and manage the objects and products they own that are recorded in the decentralized register, whether that be a watch, containers or a palette on its journey between a distributor and a shop.

To illustrate the application’s potential, Clément Bergé-Lefranc, co-founder of Ledgys, uses the example of a company producing luxury items: “a brand that produces 1,000 individually-numbered bags can create a digital asset for each one that is included in the blockchain. These assets will accompany the product throughout its whole life cycle.” From conception to destruction, including distribution and resale from individual to individual, the digital asset will follow the bag with which it is associated through all phases. Each time the bag moves from one actor of the chain to another, the transaction of the digital asset is recorded in the blockchain, proving that the exchange really took place.

“This transaction is approved by thousands of computers, and its certification is more secure than traditional financial exchanges”, claims Clément Bergé-Lefranc. The blockchain basically submits each transaction to be validated by other users of the technology. The proof of the exchange is therefore certified by all the other participants, and is then recorded alongside all other transactions that have been carried out. It is impossible to then go back and alter this information.

Also read on I’MTech: What is a blockchain?

With Ownest, users have easy access to information on the assets of the products they own. The app allows users to transfer a title to another person in a matter of seconds. There are a whole host of advantages for the consumer. In the example of a bag made by a luxury brand, the asset certifies that the product does indeed come from the manufacturer’s own workshops and that it is not a fake. Should they wish to then resell the bag to an individual, they can prove that the object is authentic. It also solves problems at customs when returning from a journey.

Monitoring the product throughout its life cycle allows businesses to better understand the secondary market. “A collector of luxury bags who only buys second-hand products is totally invisible to a brand”, highlights the Ledgys co-founder. “If they want, the collector can make themselves known to the brand and prove that they own the items, meaning that the brand can offer advantages such as a limited-edition item.” Going beyond customer relations, the blockchain is above all one of the quickest and most effective ways to manage the state of stocks. The company can see in real time which products have been handed over to distributors and which have just been stored in the warehouse. Instead of signing delivery slips on the distribution platforms, a simple transfer of the digital asset from the deliverer to the receiver will suffice. This perspective has also attracted one of the world leaders in distribution, who is now a client of Ledgys, hoping to improve the traceability of their packaging.

“These really are concrete examples of what blockchain technology can do, and it’s what we would like to offer”, Clément Bergé-Lefranc declares enthusiastically. The start-up will present these examples of how it can be used at the VivaTechnology exhibition in Paris from June 15-17. Ledgys is committed to their mission of popularizing the use of the blockchain, and collaborating with another start-up for the event: Inwibe and the app “Wibe me up”. Together, they will offer all participants the chance to vote for the best start-ups. By using blockchain technology to certify the votes, they can ensure a transparent census of the public’s favorites.

Imagining how blockchain might be used in 10 years’ time

As well as developing Ownest, the Ledgys team is working on a more long-term project. “One of our objectives is to build marketplaces for the exchange of data using blockchain technology” presents Clément Bergé-Lefranc. Using Ethereum, the solution allows people to buy and sell data with blockchain. They still have a while to wait however before seeing such marketplaces emerge. “There are technical elements that still need to be resolved for this to be possible and optimized on Ethereum”, admits the Ledgys’ co-founder. “However, we are already building on what the blockchain will become, and its use in a few years’ time.”

In the meantime, the start-up is working on a proof of concept in developing countries, in collaboration with “The Heart Fund”, a foundation devoted to treating heart disease. The UN-accredited project aims to establish a secure and universal medical file for each patient. Blockchain technology will allow health-related data to be certified and accessible. “The aim is that with time, we will promote the proper use of patients’ medical data”, announces Clément Bergé-Lefranc. By authorizing access for professionals in the medical sector in a transparent and secure way, the quality of healthcare in countries where medical attention is less reliable can be improved. Again, this is an example of Ledgys’ desire to use blockchain not just to fulfil fantasies, but also to resolve concrete problems.

The original version of this article was published on the ParisTech Entrepreneurs incubator website.

 

Data&Musée, cultural institutions

Data&Musée – Developing data science for cultural institutions

Télécom SudParis and Télécom ParisTech are taking part in Data&Musée, a collaborative project led by Orpheo, launched on September 27, 2017. The project’s aim is to provide a single, open platform for data from cultural institutions in order to develop analysis and forecasting tools to guide them in developing strategies and expanding their activities.

 

Data science is a recent scientific discipline concerned with extracting information, analyses or forecasts from a large quantity of data. It is now widely used in many different industries from energy and transport to the healthcare sector.

However, this discipline has not yet become a part of French cultural institutions’ practices. Though institutions collect their data on an individual level, until now there had been no initiative to aggregate and analyze all the data from French museums and monuments. And yet, gathering this data could provide valuable information for institutions and visitors alike, whether to establish analyses of cultural products in France, measure institutions’ performance or provide visitors with helpful recommendations for museums and monuments to visit.

The Data&Musée project will serve as a testing ground for exploring the potential of data analysis for cultural institutions and determining how this data can help institutions grow. The project is led by the Orpheo group, a provider of guide systems (audio-guide systems, multimedia guides, software etc.) for cultural and tourist sites, and has brought together researchers and a team of companies specialized in data analysis such as Tech4TeamKernixMyOrpheo. The Centre des Monuments Nationaux, an institution which groups together nearly 100 monuments, and Paris Musées, an organization which incorporates 14 museums in Paris, have agreed to test the project on their sites.

A single, open platform for centralizing data

The Data&Musée project strives to usher museums into the data age by grouping together a great number of cultural institutions on Teralab, IMT and GENES’s shared data platform. “This platform provides a neutral, secure and sovereign storage hosting space. The data will be hosted on the IMT Lille Douai site in France,” explains Antoine Garnier, the head of the project at IMT. “Teralab can host sensitive data in accordance with current regulations and is already recognized as a trustworthy tool.

In addition, highly sensitive data can be anonymized if necessary. The project could enlist the help of Lamane, a startup specializing in these technical issues, which was created through IMT Atlantique incubators.

Previously-collected individual data from each institution, such as ticketing data or web site traffic, will be combined with new sources collected by Data&Musée and created by visitors using a smart guestbook (currently being developed by the corporate partner GuestViews), social media analysis and an indoor geolocation system.

Orpheo seeks to enhance the visitor journey but is not certain whether it should be up to the visitor or carried out automatically,” explains Nel Samama, whose research laboratory at Télécom SudParis is working with Orpheo on the geolocation aspect. “Analyzing flows in a fully automatic way means using radio or optical techniques, which function correctly in demonstration mode but are unreliable in real use. Having the visitor participate in this loop would simplify it tremendously.

Developing tools for indications, forecasting and recommendations

Based on an analysis of this data, the goal is to develop performance indicators for museums and build tools for personalizing the visitor experience.

Other project partners including Reciproque, a company that provides engineering services for cultural institutions, and the UNESCO ITEN chair (Innovation, Transmission and Digital Publishing), will use the data collected to work on modeling aesthetic taste in order to determine typical visitor profiles and appropriate recommendations for content based on these profiles. This tool will therefore increase visitors’ awareness of the rich offerings of French cultural institutions and therefore boost the tourism industry. Jean-Claude Moissinac, a research professor at Télécom ParisTech, is working on this aspect of the project in partnership with Reciproque. “I’m especially interested in data semantics, or web semantics,” explains the researcher. “The idea is to index all the data collected in a homogenous way, then use it to make a graph in order to interlink the information. We can then infer groups, which may be works or users. After that, we use this knowledge to propose different paths.”

The project plans to set up an interface through which partner institutions may view their regional attendance, visitor seasonality, and segmentation compared to other institutions with similar themes. Performance indicators will also be developed for the museums. The various data collected will be used to develop analytical and predictive models for visiting cultural sites in France and for providing these institutions with recommendations to help them determine strategies for expanding their activities.

With a subscription or contribution system, this structured data could eventually be transmitted to institutions that do not produce data or to third parties with the consent of institutions and users. A business model could therefore emerge, allowing Data&Musée to live on beyond the duration of the project.

Project supported by Cap Digital and Imaginove, with funding from Bpifrance and Région Île-de-France.