Web open standards for linked data and knowledge graphs as enablers of EU digital sovereignty
ENDORSE Keynote by Fabien GANDON, 19/03/2021
https://op.europa.eu/en/web/endorse
Overview of the Research in Wimmics 2018Fabien Gandon
The WIMMICS team conducts multidisciplinary research at the intersection of natural and artificial intelligence on the semantic web. Their work involves modeling user and community data using typed graphs to enable formal analysis and the development of applications for knowledge sharing communities. Their methods include user modeling, knowledge extraction from social media, and developing question answering and serious game applications using linked open data.
This document provides a summary of the 12th European Semantic Web Conference (ESWC) that took place from May 31st to June 4th, 2015 in Portoroz, Slovenia. It outlines key details about the conference including the number of registered participants, program details such as the number of paper submissions and accepted papers by track, and highlights of the keynote speakers and events during the conference.
The document discusses the relationship between web science and artificial intelligence (AI). It makes three key points:
1) AI can be used for the web, such as using connectionist AI for clustering, recommending, classifying and indexing web data, and symbolic AI for structured querying, reasoning and interoperating.
2) The web can be used for AI, such as providing large datasets and corpus for connectionist AI training, and standards for symbolic AI knowledge representation.
3) There is an intersection between web science and AI, as the web and AI influence each other, which is an important topic for web science.
This course is a quick overview of the fundamentals of graph databases and graph queries, with a focus on RDF and SPARQL. It includes both simple and challenging hands-on exercises to practice and test your understanding.
The material for this course can be downloaded form the following link: https://github.com/paolo7/Introduction-to-Graph-Databases
ESWC 2015 Closing and "General Chair's minute of Madness"Fabien Gandon
The document summarizes the closing speech of the 12th European Semantic Web Conference held from May 31st to June 4th, 2015 in Portoroz, Slovenia. It recognizes award winners in various categories including best papers, challenges, and demos. It also announces details about the upcoming ESWC 2015 summer school and ESWC 2016 conference in Crete. The speaker emphasizes that semantic web technologies can effectively handle large volumes of multilingual data, optimize queries and reasoning, support new devices and applications, and enable predictive and collaborative capabilities.
One Web of pages, One Web of peoples, One Web of Services, One Web of Data, O...Fabien Gandon
Keynote Fabien GANDON, at WIM2016: One Web of pages, One Web of peoples, One Web of Services, One Web of Data, One Web of Things…and with the Semantic Web bind them.
On the many graphs of the Web and the interest of adding their missing links. Fabien Gandon
The document discusses research on analyzing and modeling typed graphs to formalize and implement social semantic web applications for epistemic communities. It describes challenges in bridging social semantics and formal semantics on the web using typed graphs to represent web data. It also summarizes previous work presented at ICCS on RDF and conceptual graphs and outlines the multi-disciplinary WIMMICS team researching various aspects of linked data, semantic web, communities and more.
JURIX talk on representing and reasoning on the deontic aspects of normative rules relying only on standard Semantic Web languages.
The corresponding paper is at https://hal.inria.fr/hal-01643769v1
bridging formal semantics and social semantics on the webFabien Gandon
The document summarizes research on bridging formal semantics and social semantics on the web. It discusses:
1) The Wimmics research team which studies web-instrumented machine interactions, communities, and semantics using a multidisciplinary approach and typed graphs.
2) The challenge of analyzing, modeling, and formalizing social semantic web applications for communities by combining formal semantics and social semantics.
3) Examples of past work that have structured folksonomies, combined metric spaces for tags, and analyzed sociograms and social networks.
Wimmics Research Team 2015 Activity ReportFabien Gandon
Extract of the activity report of the Wimmics joint research team between Inria Sophia Antipolis - Méditerranée and I3S (CNRS and Université Nice Sophia Antipolis). Wimmics stands for web-instrumented man-machine interactions, communities and semantics. The team focuses on bridging social semantics and formal semantics on the web.
This document discusses linking open data with Drupal. It begins with an introduction to open data and the semantic web. It explains how to transform open data into linked data using ontologies and semantic metadata. Several Drupal modules are presented for importing, publishing, and querying linked data. The document concludes by proposing a hackathon where participants could consume, publish, and build applications with linked open government data and the Drupal framework.
on the ontological necessity of the multidisciplinary development of the webFabien Gandon
Talk on the ontological necessity of the multidisciplinary development of the web at the panel CLOSER/WEBIST 2014 on "social, political and economic implications of cloud and web"
How to Build Linked Data Sites with Drupal 7 and RDFascorlosquet
Slides of the tutorial Stéphane Corlosquet, Lin Clark and Alexandre Passant presented at SemTech 2010 in San Francisco http://semtech2010.semanticuniverse.com/sessionPop.cfm?confid=42& proposalid=2889
Creating knowledge out of interlinked dataSören Auer
This document discusses creating knowledge from interlinked data. It notes that while reasoning over large datasets does not currently scale well, linked data approaches are more feasible as they allow for incremental improvement. The document outlines the linked data lifecycle including extraction, storage and querying, authoring, linking, and enrichment of semantic data. It provides examples of projects that extract, store, author and link diverse datasets including DBpedia, LinkedGeoData, and statistical data. Challenges discussed include improving query performance, developing standardized interfaces, and increasing the amount of interlinking between datasets.
From Open Linked Data towards an Ecosystem of Interlinked KnowledgeSören Auer
This document discusses the development of linked open data and its potential to create an ecosystem of interlinked knowledge. It outlines achievements in extending the web with structured data and the growth of an open research community. However, it also identifies challenges regarding coherence, quality, performance and usability that must be addressed for linked data to reach its full potential as a global platform for knowledge integration. The document proposes that addressing these issues could ultimately lead to an ecosystem of interlinked knowledge on the semantic web.
From the Semantic Web to the Web of Data: ten years of linking upDavide Palmisano
This document discusses the concepts and technologies behind the Semantic Web. It describes how RDF, RDF Schema, and OWL allow structured data and relationships to be represented and shared across the web. It also discusses tools for working with semantic data in Java, such as Jena, Sesame, and Any23 for extracting and working with RDF. The document provides examples of representing data and relationships in RDF and querying semantic data with SPARQL.
The document provides an introduction to Prof. Dr. Sören Auer and his background in knowledge graphs. It discusses his current role as a professor and director focusing on organizing research data using knowledge graphs. It also briefly outlines some of his past roles and major scientific contributions in the areas of technology platforms, funding acquisition, and strategic projects related to knowledge graphs.
The document summarizes a lecture on the Semantic Web, with the purpose of introducing key concepts and some advanced ideas that would be useful for graduating seniors. The lecture covers the evolution of the Semantic Web from its origins in 1989, and explains core concepts like RDF, SPARQL, Linked Data, and how they enable applications like mashups and data visualization. Some advanced ideas discussed include ontology, inference, reasoning and machine learning.
Das Semantische Daten Web für UnternehmenSören Auer
This document summarizes the vision, technology, and applications of the Semantic Data Web for businesses. It discusses how the Semantic Web can help solve problems of searching for complex information across different data sources by complementing text on web pages with structured linked open data. It provides overviews of RDF standards, vocabularies, and technologies like SPARQL and OntoWiki that allow creating and managing structured knowledge bases. It also presents examples like DBpedia that extract structured data from Wikipedia and make it available on the web as linked open data.
Learning Multilingual Semantics from Big Data on the WebGerard de Melo
This document summarizes Gerard de Melo's presentation on learning multilingual semantics from big data on the web. It discusses how lexical and taxonomic knowledge can be extracted at large scale from online resources like Wiktionary, Wikipedia, and WordNet. Methods are presented for merging structured data like knowledge graphs and integrating taxonomies across languages using techniques like linear program relaxation and belief propagation. The goal is to build large yet reasonably clean multilingual knowledge bases to power applications in areas like semantic search and the digital humanities.
This document summarizes recent approaches to web data management including Fusion Tables, XML, and Linked Open Data (LOD). It discusses properties of web data like lack of schema, volatility, and scale. LOD uses RDF, global identifiers (URIs), and data links to query and integrate data from multiple sources while maintaining source autonomy. The LOD cloud has grown rapidly, currently consisting of over 3000 datasets with more than 84 billion triples.
This document discusses live social semantics and monitoring online communities. It presents approaches to integrate physical presence data from RFID sensors with online social network and semantic web data. This would allow for semantic user profiling, logging face-to-face contacts, and analyzing how online and offline social networks converge. The live social semantics architecture extracts and links social media data to semantic web data sources and stores it in a triple store for analysis and applications like social network browsing.
The document discusses beautifying data in the real world. It describes how much data exists on the internet, which is estimated to reach nearly 1,000 exabytes by 2015. It also discusses open notebook science, crowdsourcing data, and challenges with real world data like noise and barriers to presentation. Unique identifiers for chemicals and options for analyzing data are examined. The document proposes using semantic web technologies like RDF and SPARQL to build knowledge from beautified data and create non-obvious relationships. It demonstrates visualizing data through services like Google Docs and Second Life.
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
The document discusses the ISICIL project which aims to develop an enterprise social networking platform called ISICIL.inria.fr. The platform will integrate requirement analysis methods and allow for functions like business intelligence, monitoring communities of interest and experts. Examples are provided of challenges addressed by the platform like assisting the structuring of folksonomies, detecting relationships between tags, and analyzing social networks and graphs within organizations. The document outlines the ISICIL project which combines formal semantics, social semantics and graphs to support corporate intelligence applications.
DISIT Lab overview: smart city, big data, semantic computing, cloudPaolo Nesi
Smart City
• Projects: http://www.disit.org/5501
– Sii-Mobility, http://www.sii-mobility.org
– Service Map: http://servicemap.disit.org
– Social Innovation: Coll@bora http://www.disit.org/5479
– Navigation Indoor/outdoor: Mobile Emergency http://www.disit.org/5404
– Mobility and Transport: TRACE-IT, RAISSS, TESYSRAIL
• Tools: http://www.disit.org/5489
– Data gathering, data mining and reconciliation
– Data reasoning, deduction, prediction
– Smart city ontology and reasoning tools
– Service analysis and recommendations
– Autonomous train operator, train signaling
– Risk analysis, decision support systems
– Mobile Applications
Data Analytics - Big data
• Projects: http://www.disit.org/5501
– Linked Open Graph: http://LOG.disit.org
– Sii-Mobility, http://www.sii-mobility.org
– Service on a number of projects
• Tools: http://www.disit.org/5489
– Open data and Linked Open Data
– LOG LOD service and tools
– Data mining and reconciliation
– Data reasoning, deduction, prediction, decision support
– SN Analysis and recommendations
– User behavior monitoring and analysis
Smart Cloud - Computing
• Projects: http://www.disit.org/5501
– ICARO: http://www.disit.org/5482
– Cloud ontology: http://www.disit.org/5604
– Cloud simulator:
– Smart Cloud: http://www.disit.org/6544
• Tools: http://www.disit.org/5489
– Cloud Monitoring
– Smart Cloud Engine and reasoner,
– Service Level Analyzer and control
– Configuration analysis and checker
– Cloud Simulation
Text and Web Mining
• Projects: http://www.disit.org/5501
– OSIM: http://www.disit.org/5482
– SACVAR: http://www.disit.org/5604
– Blog/Twitter Vigilance
• Tools: http://www.disit.org/5489
– Text and web mining, Natural Language Processing
– Service localization
– Web Crawling
– Competence analysis
– Blog Vigiliance, sentiment analysis
Social Media and e-Learning
• Projects: http://www.disit.org/5501
– ECLAP, http://www.eclap.eu
– ApreToscana: http://www.apretoscana.org
– Others: AXMEDIS, VARIAZIONI, SMNET, etc.
– Samsung Smart TV: http://www.disit.org/6534
• Tools: http://www.disit.org/5489
– XLMS, Cross Media Learning System
– IPR and content protection and distribution
– Mobile and SmartTv Applications
– Suggestions and recommendations
– Matchmaking solutions
– Media Tools for cross media content
Mobile Computing
• Projects:
– ECLAP: http://www.eclap.eu
– Mobile Medicine: http://mobmed.axmedis.org
– Mobile Emergency: http://www.disit.org/5500
– Smart City, FODD 2015: http://www.disit.org/6593
– Resolute: Mobiles as sensors
• Tools and support:
– Content distribution: e-learning
– Integrated Indoor/outdoor navigation
– User networking and collaboration
– Service localization
– Smart city and services
– OS: iOS, Android, Windows Phone, etc.
– Tech: IOT, iBeacoms, NFC, QR, ….
One Web of pages, One Web of peoples, One Web of Services, One Web of Data, O...Fabien Gandon
Keynote Fabien GANDON, at WIM2016: One Web of pages, One Web of peoples, One Web of Services, One Web of Data, One Web of Things…and with the Semantic Web bind them.
On the many graphs of the Web and the interest of adding their missing links. Fabien Gandon
The document discusses research on analyzing and modeling typed graphs to formalize and implement social semantic web applications for epistemic communities. It describes challenges in bridging social semantics and formal semantics on the web using typed graphs to represent web data. It also summarizes previous work presented at ICCS on RDF and conceptual graphs and outlines the multi-disciplinary WIMMICS team researching various aspects of linked data, semantic web, communities and more.
JURIX talk on representing and reasoning on the deontic aspects of normative rules relying only on standard Semantic Web languages.
The corresponding paper is at https://hal.inria.fr/hal-01643769v1
bridging formal semantics and social semantics on the webFabien Gandon
The document summarizes research on bridging formal semantics and social semantics on the web. It discusses:
1) The Wimmics research team which studies web-instrumented machine interactions, communities, and semantics using a multidisciplinary approach and typed graphs.
2) The challenge of analyzing, modeling, and formalizing social semantic web applications for communities by combining formal semantics and social semantics.
3) Examples of past work that have structured folksonomies, combined metric spaces for tags, and analyzed sociograms and social networks.
Wimmics Research Team 2015 Activity ReportFabien Gandon
Extract of the activity report of the Wimmics joint research team between Inria Sophia Antipolis - Méditerranée and I3S (CNRS and Université Nice Sophia Antipolis). Wimmics stands for web-instrumented man-machine interactions, communities and semantics. The team focuses on bridging social semantics and formal semantics on the web.
This document discusses linking open data with Drupal. It begins with an introduction to open data and the semantic web. It explains how to transform open data into linked data using ontologies and semantic metadata. Several Drupal modules are presented for importing, publishing, and querying linked data. The document concludes by proposing a hackathon where participants could consume, publish, and build applications with linked open government data and the Drupal framework.
on the ontological necessity of the multidisciplinary development of the webFabien Gandon
Talk on the ontological necessity of the multidisciplinary development of the web at the panel CLOSER/WEBIST 2014 on "social, political and economic implications of cloud and web"
How to Build Linked Data Sites with Drupal 7 and RDFascorlosquet
Slides of the tutorial Stéphane Corlosquet, Lin Clark and Alexandre Passant presented at SemTech 2010 in San Francisco http://semtech2010.semanticuniverse.com/sessionPop.cfm?confid=42& proposalid=2889
Creating knowledge out of interlinked dataSören Auer
This document discusses creating knowledge from interlinked data. It notes that while reasoning over large datasets does not currently scale well, linked data approaches are more feasible as they allow for incremental improvement. The document outlines the linked data lifecycle including extraction, storage and querying, authoring, linking, and enrichment of semantic data. It provides examples of projects that extract, store, author and link diverse datasets including DBpedia, LinkedGeoData, and statistical data. Challenges discussed include improving query performance, developing standardized interfaces, and increasing the amount of interlinking between datasets.
From Open Linked Data towards an Ecosystem of Interlinked KnowledgeSören Auer
This document discusses the development of linked open data and its potential to create an ecosystem of interlinked knowledge. It outlines achievements in extending the web with structured data and the growth of an open research community. However, it also identifies challenges regarding coherence, quality, performance and usability that must be addressed for linked data to reach its full potential as a global platform for knowledge integration. The document proposes that addressing these issues could ultimately lead to an ecosystem of interlinked knowledge on the semantic web.
From the Semantic Web to the Web of Data: ten years of linking upDavide Palmisano
This document discusses the concepts and technologies behind the Semantic Web. It describes how RDF, RDF Schema, and OWL allow structured data and relationships to be represented and shared across the web. It also discusses tools for working with semantic data in Java, such as Jena, Sesame, and Any23 for extracting and working with RDF. The document provides examples of representing data and relationships in RDF and querying semantic data with SPARQL.
The document provides an introduction to Prof. Dr. Sören Auer and his background in knowledge graphs. It discusses his current role as a professor and director focusing on organizing research data using knowledge graphs. It also briefly outlines some of his past roles and major scientific contributions in the areas of technology platforms, funding acquisition, and strategic projects related to knowledge graphs.
The document summarizes a lecture on the Semantic Web, with the purpose of introducing key concepts and some advanced ideas that would be useful for graduating seniors. The lecture covers the evolution of the Semantic Web from its origins in 1989, and explains core concepts like RDF, SPARQL, Linked Data, and how they enable applications like mashups and data visualization. Some advanced ideas discussed include ontology, inference, reasoning and machine learning.
Das Semantische Daten Web für UnternehmenSören Auer
This document summarizes the vision, technology, and applications of the Semantic Data Web for businesses. It discusses how the Semantic Web can help solve problems of searching for complex information across different data sources by complementing text on web pages with structured linked open data. It provides overviews of RDF standards, vocabularies, and technologies like SPARQL and OntoWiki that allow creating and managing structured knowledge bases. It also presents examples like DBpedia that extract structured data from Wikipedia and make it available on the web as linked open data.
Learning Multilingual Semantics from Big Data on the WebGerard de Melo
This document summarizes Gerard de Melo's presentation on learning multilingual semantics from big data on the web. It discusses how lexical and taxonomic knowledge can be extracted at large scale from online resources like Wiktionary, Wikipedia, and WordNet. Methods are presented for merging structured data like knowledge graphs and integrating taxonomies across languages using techniques like linear program relaxation and belief propagation. The goal is to build large yet reasonably clean multilingual knowledge bases to power applications in areas like semantic search and the digital humanities.
This document summarizes recent approaches to web data management including Fusion Tables, XML, and Linked Open Data (LOD). It discusses properties of web data like lack of schema, volatility, and scale. LOD uses RDF, global identifiers (URIs), and data links to query and integrate data from multiple sources while maintaining source autonomy. The LOD cloud has grown rapidly, currently consisting of over 3000 datasets with more than 84 billion triples.
This document discusses live social semantics and monitoring online communities. It presents approaches to integrate physical presence data from RFID sensors with online social network and semantic web data. This would allow for semantic user profiling, logging face-to-face contacts, and analyzing how online and offline social networks converge. The live social semantics architecture extracts and links social media data to semantic web data sources and stores it in a triple store for analysis and applications like social network browsing.
The document discusses beautifying data in the real world. It describes how much data exists on the internet, which is estimated to reach nearly 1,000 exabytes by 2015. It also discusses open notebook science, crowdsourcing data, and challenges with real world data like noise and barriers to presentation. Unique identifiers for chemicals and options for analyzing data are examined. The document proposes using semantic web technologies like RDF and SPARQL to build knowledge from beautified data and create non-obvious relationships. It demonstrates visualizing data through services like Google Docs and Second Life.
Talk about Exploring the Semantic Web, and particularly Linked Data, and the Rhizomer approach. Presented August 14th 2012 at the SRI AIC Seminar Series, Menlo Park, CA
The document discusses the ISICIL project which aims to develop an enterprise social networking platform called ISICIL.inria.fr. The platform will integrate requirement analysis methods and allow for functions like business intelligence, monitoring communities of interest and experts. Examples are provided of challenges addressed by the platform like assisting the structuring of folksonomies, detecting relationships between tags, and analyzing social networks and graphs within organizations. The document outlines the ISICIL project which combines formal semantics, social semantics and graphs to support corporate intelligence applications.
DISIT Lab overview: smart city, big data, semantic computing, cloudPaolo Nesi
Smart City
• Projects: http://www.disit.org/5501
– Sii-Mobility, http://www.sii-mobility.org
– Service Map: http://servicemap.disit.org
– Social Innovation: Coll@bora http://www.disit.org/5479
– Navigation Indoor/outdoor: Mobile Emergency http://www.disit.org/5404
– Mobility and Transport: TRACE-IT, RAISSS, TESYSRAIL
• Tools: http://www.disit.org/5489
– Data gathering, data mining and reconciliation
– Data reasoning, deduction, prediction
– Smart city ontology and reasoning tools
– Service analysis and recommendations
– Autonomous train operator, train signaling
– Risk analysis, decision support systems
– Mobile Applications
Data Analytics - Big data
• Projects: http://www.disit.org/5501
– Linked Open Graph: http://LOG.disit.org
– Sii-Mobility, http://www.sii-mobility.org
– Service on a number of projects
• Tools: http://www.disit.org/5489
– Open data and Linked Open Data
– LOG LOD service and tools
– Data mining and reconciliation
– Data reasoning, deduction, prediction, decision support
– SN Analysis and recommendations
– User behavior monitoring and analysis
Smart Cloud - Computing
• Projects: http://www.disit.org/5501
– ICARO: http://www.disit.org/5482
– Cloud ontology: http://www.disit.org/5604
– Cloud simulator:
– Smart Cloud: http://www.disit.org/6544
• Tools: http://www.disit.org/5489
– Cloud Monitoring
– Smart Cloud Engine and reasoner,
– Service Level Analyzer and control
– Configuration analysis and checker
– Cloud Simulation
Text and Web Mining
• Projects: http://www.disit.org/5501
– OSIM: http://www.disit.org/5482
– SACVAR: http://www.disit.org/5604
– Blog/Twitter Vigilance
• Tools: http://www.disit.org/5489
– Text and web mining, Natural Language Processing
– Service localization
– Web Crawling
– Competence analysis
– Blog Vigiliance, sentiment analysis
Social Media and e-Learning
• Projects: http://www.disit.org/5501
– ECLAP, http://www.eclap.eu
– ApreToscana: http://www.apretoscana.org
– Others: AXMEDIS, VARIAZIONI, SMNET, etc.
– Samsung Smart TV: http://www.disit.org/6534
• Tools: http://www.disit.org/5489
– XLMS, Cross Media Learning System
– IPR and content protection and distribution
– Mobile and SmartTv Applications
– Suggestions and recommendations
– Matchmaking solutions
– Media Tools for cross media content
Mobile Computing
• Projects:
– ECLAP: http://www.eclap.eu
– Mobile Medicine: http://mobmed.axmedis.org
– Mobile Emergency: http://www.disit.org/5500
– Smart City, FODD 2015: http://www.disit.org/6593
– Resolute: Mobiles as sensors
• Tools and support:
– Content distribution: e-learning
– Integrated Indoor/outdoor navigation
– User networking and collaboration
– Service localization
– Smart city and services
– OS: iOS, Android, Windows Phone, etc.
– Tech: IOT, iBeacoms, NFC, QR, ….
The document discusses the context and goals of e-science and e-research, including enabling collaboration through distributed computation and data sharing. It provides examples of UK e-science initiatives like national centers and describes the role of the National e-Science Centre in Glasgow in supporting various projects through grid computing resources and expertise. Security challenges around authentication, authorization and auditing are discussed in the context of user-oriented and federated approaches.
This document discusses computational workflows and FAIR principles. It begins by providing background on computational workflows and their increasing importance. It then discusses challenges around finding, accessing, and sharing workflows. Next, it explores how applying FAIR principles to workflows could help address these challenges by making workflows and their associated objects findable, accessible, interoperable, and reusable. This includes discussing applying metadata standards, using persistent identifiers, and developing principles for FAIR workflows and FAIR software. The document concludes by examining the roles and responsibilities of different stakeholders in working towards FAIR workflows.
The document provides an overview of the Dublinked Technology Workshop held on December 15th, 2011. It includes presentations on transportation data, spatial web services, linked data, and semantic data description. Breakout sessions covered topics like data publishing, discovery, web services, and advanced functions. The workshop aimed to address challenges around sharing digital data between organizations and discussed technical requirements and tools to support open government data platforms.
RDMkit, a Research Data Management Toolkit. Built by the Community for the ...Carole Goble
https://datascience.nih.gov/news/march-data-sharing-and-reuse-seminar 11 March 2022
Starting in 2023, the US National Institutes of Health (NIH) will require institutes and researchers receiving funding to include a Data Management Plan (DMP) in their grant applications, including the making their data publicly available. Similar mandates are already in place in Europe, for example a DMP is mandatory in Horizon Europe projects involving data.
Policy is one thing - practice is quite another. How do we provide the necessary information, guidance and advice for our bioscientists, researchers, data stewards and project managers? There are numerous repositories and standards. Which is best? What are the challenges at each step of the data lifecycle? How should different types of data? What tools are available? Research Data Management advice is often too general to be useful and specific information is fragmented and hard to find.
ELIXIR, the pan-national European Research Infrastructure for Life Science data, aims to enable research projects to operate “FAIR data first”. ELIXIR supports researchers across their whole RDM lifecycle, navigating the complexity of a data ecosystem that bridges from local cyberinfrastructures to pan-national archives and across bio-domains.
The ELIXIR RDMkit (https://rdmkit.elixir-europe.org (link is external)) is a toolkit built by the biosciences community, for the biosciences community to provide the RDM information they need. It is a framework for advice and best practice for RDM and acts as a hub of RDM information, with links to tool registries, training materials, standards, and databases, and to services that offer deeper knowledge for DMP planning and FAIR-ification practices.
Launched in March 2021, over 120 contributors have provided nearly 100 pages of content and links to more than 300 tools. Content covers the data lifecycle and specialized domains in biology, national considerations and examples of “tool assemblies” developed to support RDM. It has been accessed by over 123 countries, and the top of the access list is … the United States.
The RDMkit is already a recommended resource of the European Commission. The platform, editorial, and contributor methods helped build a specialized sister toolkit for infectious diseases as part of the recently launched BY-COVID project. The toolkit’s platform is the simplest we could manage - built on plain GitHub - and the whole development and contribution approach tailored to be as lightweight and sustainable as possible.
In this talk, Carole and Frederik will present the RDMkit; aims and context, content, community management, how folks can contribute, and our future plans and potential prospects for trans-Atlantic cooperation.
Data policy must be partnered with data practice. Our researchers need to be the best informed in order to meet these new data management and data sharing mandates.
COMBINE 2019, EU-STANDS4PM, Heidelberg, Germany 18 July 2019
FAIR: Findable Accessable Interoperable Reusable. The “FAIR Principles” for research data, software, computational workflows, scripts, or any other kind of Research Object one can think of, is now a mantra; a method; a meme; a myth; a mystery. FAIR is about supporting and tracking the flow and availability of data across research organisations and the portability and sustainability of processing methods to enable transparent and reproducible results. All this is within the context of a bottom up society of collaborating (or burdened?) scientists, a top down collective of compliance-focused funders and policy makers and an in-the-middle posse of e-infrastructure providers.
Making the FAIR principles a reality is tricky. They are aspirations not standards. They are multi-dimensional and dependent on context such as the sensitivity and availability of the data and methods. We already see a jungle of projects, initiatives and programmes wrestling with the challenges. FAIR efforts have particularly focused on the “last mile” – “FAIRifying” destination community archive repositories and measuring their “compliance” to FAIR metrics (or less controversially “indicators”). But what about FAIR at the first mile, at source and how do we help Alice and Bob with their (secure) data management? If we tackle the FAIR first and last mile, what about the FAIR middle? What about FAIR beyond just data – like exchanging and reusing pipelines for precision medicine?
Since 2008 the FAIRDOM collaboration [1] has worked on FAIR asset management and the development of a FAIR asset Commons for multi-partner researcher projects [2], initially in the Systems Biology field. Since 2016 we have been working with the BioCompute Object Partnership [3] on standardising computational records of HTS precision medicine pipelines.
So, using our FAIRDOM and BioCompute Object binoculars let’s go on a FAIR safari! Let’s peruse the ecosystem, observe the different herds and reflect what where we are for FAIR personalised medicine.
References
[1] http://www.fair-dom.org
[2] http://www.fairdomhub.org
[3] http://www.biocomputeobject.org
FAIRy stories: the FAIR Data principles in theory and in practiceCarole Goble
https://ucsb.zoom.us/meeting/register/tZYod-ippz4pHtaJ0d3ERPIFy2QIvKqjwpXR
FAIRy stories: the FAIR Data principles in theory and in practice
The ‘FAIR Guiding Principles for scientific data management and stewardship’ [1] launched a global dialogue within research and policy communities and started a journey to wider accessibility and reusability of data and preparedness for automation-readiness (I am one of the army of authors). Over the past 5 years FAIR has become a movement, a mantra and a methodology for scientific research and increasingly in the commercial and public sector. FAIR is now part of NIH, European Commission and OECD policy. But just figuring out what the FAIR principles really mean and how we implement them has proved more challenging than one might have guessed. To quote the novelist Rick Riordan “Fairness does not mean everyone gets the same. Fairness means everyone gets what they need”.
As a data infrastructure wrangler I lead and participate in projects implementing forms of FAIR in pan-national European biomedical Research Infrastructures. We apply web-based industry-lead approaches like Schema.org; work with big pharma on specialised FAIRification pipelines for legacy data; promote FAIR by Design methodologies and platforms into the researcher lab; and expand the principles of FAIR beyond data to computational workflows and digital objects. Many use Linked Data approaches.
In this talk I’ll use some of these projects to shine some light on the FAIR movement. Spoiler alert: although there are technical issues, the greatest challenges are social. FAIR is a team sport. Knowledge Graphs play a role – not just as consumers of FAIR data but as active contributors. To paraphrase another novelist, “It is a truth universally acknowledged that a Knowledge Graph must be in want of FAIR data.”
[1] Wilkinson, M., Dumontier, M., Aalbersberg, I. et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci Data 3, 160018 (2016). https://doi.org/10.1038/sdata.2016.18
Data management plans – EUDAT Best practices and case study | www.eudat.euEUDAT
| www.eudat.eu | Presentation given by Stéphane Coutin during the PRACE 2017 Spring School joint training event with the EU H2020 VI-SEEM project (https://vi-seem.eu/) organised by CaSToRC at The Cyprus Institute. Science and more specifically projects using HPC is facing a digital data explosion. Instruments and simulations are producing more and more volume; data can be shared, mined, cited, preserved… They are a great asset, but they are facing risks: we can miss storage, we can lose them, they can be misused,… To start this session, we will review why it is important to manage research data and how to do this by maintaining a Data Management Plan. This will be based on the best practices from EUDAT H2020 project and European Commission recommendation. During the second part we will interactively draft a DMP for a given use case.
The Linked Data Research Centre (LiDRC) is a new effort within DERI to advance linked data research and development. The LiDRC operates across existing units and has 11 DERI researchers working with 9 international peers. Its research themes include publishing, discovery, application domains, and streamed linked data. The LiDRC contributes linked data infrastructure, provides tools and libraries, and participates in standards activities. It is calling for input on a technical report about linked data applications.
This document summarizes efforts to publish clinical quality data from health.data.gov as linked open data. It describes releasing metadata and data from the Hospital Compare project as RDF using vocabularies like VoID, FOAF and DC. Tools like Google Refine, Top Braid Composer and Virtuoso were used to transform, model and serve the data. A community of practice seeks to evolve standards and share best practices for publishing government linked data.
Ontology Building vs Data Harvesting and Cleaning for Smart-city ServicesPaolo Nesi
Presently, a very large number of public and private data sets are available around the local governments. In most cases, they are not semantically interoperable and a huge human effort is needed to create integrated ontologies and knowledge base for smart city. Smart City ontology is not yet standardized, and a lot of research work is needed to identify models that can easily support the data reconciliation, the management of the complexity and reasoning. In this paper, a system for data ingestion and reconciliation of smart cities related aspects as road graph, services available on the roads, traffic sensors etc., is proposed. The system allows managing a big volume of data coming from a variety of sources considering both static and dynamic data. These data are mapped to smart-city ontology and stored into an RDF-Store where they are available for applications via SPARQL queries to provide new services to the users. The paper presents the process adopted to produce the ontology and the knowledge base and the mechanisms adopted for the verification, reconciliation and validation. Some examples about the possible usage of the coherent knowledge base produced are also offered and are accessible from the RDF-Store and related services. The article also presented the work performed about reconciliation algorithms and their comparative assessment and selection. Keywords Smart city, knowledge base construction, reconciliation, validation and verification of knowledge base, smart city ontology, linked open graph.
Approximation and Self-Organisation on the Web of DataKathrin Dentler
This document discusses using computational intelligence techniques like evolutionary computing and collective intelligence to handle challenges posed by the growing Web of Data. It describes how these techniques can provide adaptive, scalable, and robust approaches to tasks like ontology mapping, query answering, and reasoning. Evolutionary computing is proposed for optimization problems, while collective intelligence approaches may enable emergent behaviors from decentralized data flows and reasoning. While computational intelligence loses precision, it gains properties like adaptation, simplicity, scalability, and interactive behavior that are well-suited to the dynamic, distributed nature of the Web of Data.
The document discusses FAIR (Findable, Accessible, Interoperable, Reusable) principles for data and their implementation. It describes the origins of FAIR, defines what makes data FAIR, and discusses tools for evaluating FAIRness like FAIRsharing and FAIR metrics. It also outlines a strategy for implementing FAIR metrics in the ASIO project, including developing a bridge between ASIO and FAIRmetrics and using it to evaluate resources and ASIO's ontology network for FAIR compliance.
Make our Scientific Datasets Accessible and Interoperable on the WebFranck Michel
The presentation investigates the challenges that we must face to share scientific datasets on the Web following the Linked Open Data principles. We present the standards of the Semantic Web and investigate how they can help address those challenges. We give tips as to how to choose vocabularies to describe data and metadata, link datasets to other related datasets by making appropriate alignments, translate existing data sources to RDF and publish it on the Web as linked data.
The document summarizes a talk given by Dr. Johannes Keizer on the CIARD (Coherence in Information for Agricultural Research for development) initiative and a global infrastructure for linked open data (LOD). The CIARD initiative aims to provide open access to agricultural research by promoting standards and sharing information. It involves institutions contributing their research outputs through the CIARD RING and adopting standards. The infrastructure proposed includes distributed repositories linked through vocabularies and LOD. Tools are being developed to generate LOD and link datasets through shared concepts.
This is one out of a series of presentations which I have given during a recent trip to the United States. I will make them all public, but content does not vary a lot between some of them
The document discusses the adoption of semantic web technologies. It notes that while tools and specifications have matured, applications are now coming to the forefront. It provides examples of semantic web deployments in various domains like digital libraries, eGovernment, healthcare and by major companies. Real-world applications include data integration, intelligent portals and knowledge management systems. Overall adoption is increasing but skills and training remain obstacles.
Infrastructures Supporting Inter-disciplinary Research - Exemplars from the UK NeISSProject
Infrastructures Supporting Inter-disciplinary Research - Exemplars from the UK . Talk given by Richard Sinnott at Urban Research Infrastructure Network Workshops, Melbourne, Brisbane, Sydney, September 2010.
Walking Our Way to the Web - Fabien Gandon
The Web: Scientific Creativity, Technological Innovation and Society
XXVIII Conference on Contemporary Philosophy and Methodology of Science
9 and 10 March 2023
University of A Coruña
The prospect of Walking our Way to the Web may sound strange to contemporary readers of this article for whom the Web is omnipresent. However, the slogan of the World Wide Web Consortium (W3C) has been, for years, and remains today, to lead “the Web to its full potential” meaning we haven’t reached that potential yet, whatever it is. The first architect of the Web himself, Tim Berners-Lee, said in an interview in 2009: “The Web as I envisaged it, we have not seen it yet. The future is still so much bigger than the past”. And he is still very active, together with the W3C members and Web experts world-wide, in proposing evolutions of the Web architecture to improve its growing usages and applications. In this article we will review the path that led us to the actual Web, the shape it is taking now and the possible evolutions, good and bad, we can identify today. This will lead us to consider the distance that we witness between the initial vision and the reality of the Web today, and to reflect on the possible divergence between the potential we see in the Web and the directions it could take. Our goal in this article is to reflect on how we could walk the delicate path to the full potential of the Web, finding the missing links and avoiding the one too many links.
a shift in our research focus: from knowledge acquisition to knowledge augmen...Fabien Gandon
EKAW 2022 keynote by Fabien GANDON: "a shift in our research focus: from knowledge acquisition to knowledge augmentation"
While EKAW started in 1987 as the European Knowledge Acquisition Workshop, in 2000 it transformed into a conference where we advance knowledge engineering and modelling in general. At the time, this transition also echoed shifts of focus such as moving from the paradigm of expert systems to the more encompassing one of knowledge-based systems. Nowadays, with the current strong interest for knowledge graphs, it is important again to reaffirm that our ultimate goal is not the acquisition of bigger siloed knowledge bases but to support knowledge requisition by and for all kinds of intelligence. Knowledge without intelligence is a highly perishable resource. Intelligence without knowledge is doomed to stagnation. We will defend that intelligence and knowledge, and their evolutions, have to be considered jointly and that the Web is providing a social hypermedia to link them in all their forms. Using examples from several projects, we will suggest that, just like intelligence augmentation and amplification insist on putting humans at the center of the design of artificial intelligence methods, we should think in terms of knowledge augmentation and amplification and we should design a knowledge web to be an enabler of the futures we want.
A Never-Ending Project for Humanity Called “the Web”Fabien Gandon
The document summarizes the history and evolution of the World Wide Web. It describes early influences and concepts from scientists like Vannevar Bush and Ted Nelson. It then outlines Tim Berners-Lee's work at CERN in the late 1980s to develop the first successful implementation of the Web including creating HTML, URIs, HTTP, and the first web browser and server. The document discusses how the Web architecture was designed to be simple, extensible and "viral" which contributed to its widespread adoption. It notes ongoing mutations and future potential areas of development for the Web including mobile access, semantic web, social media, internet of things, and more. The document concludes the Web remains a "never-ending project" with ongoing
CovidOnTheWeb : covid19 linked data published on the WebFabien Gandon
The Covid-on-the-Web project aims to allow biomedical researchers to access, query and make sense of COVID-19 related literature. To do so, it adapts, combines and extends tools to process, analyze and enrich the "COVID-19 Open Research Dataset" (CORD-19) that gathers 50,000+ full-text scientific articles related to the coronaviruses. We report on the RDF dataset and software resources produced in this project by leveraging skills in knowledge representation, text, data and argument mining, as well as data visualization and exploration. The dataset comprises two main knowledge graphs describing (1) named entities mentioned in the CORD-19 corpus and linked to DBpedia, Wikidata and other BioPortal vocabularies, and (2) arguments extracted using ACTA, a tool automating the extraction and visualization of argumentative graphs, meant to help clinicians analyze clinical trials and make decisions. On top of this dataset, we provide several visualization and exploration tools based on the Corese Semantic Web platform, MGExplorer visualization library, as well as the Jupyter Notebook technology. All along this initiative, we have been engaged in discussions with healthcare and medical research institutes to align our approach with the actual needs of the biomedical community, and we have paid particular attention to comply with the open and reproducible science goals, and the FAIR principles.
from linked data & knowledge graphs to linked intelligence & intelligence graphsFabien Gandon
ISWC Vision track talk "from linked data & knowledge graphs to linked intelligence & intelligence graphs or the potential of the semantic Web to break the walls between semantic networks and computational networks"
The document discusses various tips for PhD students to establish effective relationships and communication with their supervisors. It suggests that students should clarify roles and expectations with their supervisor, gain autonomy over their project gradually, understand their supervisor's incentives and communication style, and be proactive in providing updates, setting deadlines, organizing meetings, and documenting progress. Regular communication through writing reports and keeping supervisors informed is important for successful supervision.
Retours sur le MOOC "Web Sémantique et Web de données"Fabien Gandon
Présentation des caractéristiques et résultats de la première session en 2015 du MOOC "Web Sémantique et Web de données" par Inria, Université de Nice, FUN et UNIT.
Emotions in Argumentation: an Empirical Evaluation @ IJCAI 2015Fabien Gandon
This document summarizes a study that analyzed the connection between arguments made in debates and the emotional states of the participants. The study used tools to detect emotions like facial expressions and engagement levels to analyze 12 debates involving 20 participants on various topics. It found correlations between emotions like anger and the number of attacks, as well as between engagement and the number of arguments made. The study aimed to understand how emotions relate to argumentation online to help detect when debates become uncivil or reach consensus. It released a dataset of the debates, arguments, and emotional states to allow further analysis.
Nous lisons régulièrement que le Web révolutionne notre monde et provoque des évolutions dans toutes les dimensions de notre société. Mais le Web lui-même, ses usages et la compréhension que nous en avons n’ont pas cessé d’évoluer depuis la proposition à l’origine de sa création en 1989. C’est un espace en perpétuelle recréation qui nous demande sans cesse de nouvelles explorations et reconsidérations. Ce sont certains de ces changements passés, actuels, et à venir du Web que nous allons regarder ensemble en insistant sur la complexité de cet artefact qui en fait un objet de recherches pluridisciplinaires.
On Youtube: https://youtu.be/jNjHdqS-1Ko
Data protection and security on the web, ESWC2014 PanelFabien Gandon
This document discusses security issues related to semantics and the semantic web. It notes that while semantics allow for more expressive security policies, they can also enable de-anonymizing data and drawing unwanted conclusions. The document advocates for security standards and frameworks to provide security at every level of technology deployment. It also argues that security is as much a social and human issue as a technical one, requiring sustained attention to factors like awareness, context changes, and preferences to prevent problems like social engineering.
An introduction to Semantic Web and Linked DataFabien Gandon
Here are the steps to answer this SPARQL query against the given RDF base:
1. The query asks for all ?name values where there is a triple with predicate "name" and another triple with the same subject and predicate "email".
2. In the base, _:b is the only resource that has both a "name" and "email" triple.
3. _:b has the name "Thomas".
Therefore, the only result of the query is ?name = "Thomas".
So the result of the SPARQL query is:
?name
"Thomas"
The document discusses the Semantic Web and linked data. It describes standards like RDF, RDFS, and OWL that add structure and meaning to data on the web. Triples are used to represent information that can then be queried or linked to other data to form a global graph. The principles of linked data encourage using URIs, HTTP, and content negotiation to publish and interconnect structured data on the web.
Données de la culture et culture des donnéesFabien Gandon
Présentation "Données de la culture et culture des données" ou le web sémantique et les données liées sur le web dans le domaine de la culture à l'occasion de la conférence "Transmettre la culture à l’ère du numérique" dans le programme Automne Numérique du ministère de la Culture et de la Communication.
La vidéo de la conférence est ici:
http://www.dailymotion.com/video/x17i1g6_conference-transmettre-la-culture-a-l-age-du-numerique-fabien-gandon_tech
Just-in-time: Repetitive production system in which processing and movement of materials and goods occur just as they are needed, usually in small batches
JIT is characteristic of lean production systems
JIT operates with very little “fat”
computer organization and assembly language : its about types of programming language along with variable and array description..https://www.nfciet.edu.pk/
This comprehensive Data Science course is designed to equip learners with the essential skills and knowledge required to analyze, interpret, and visualize complex data. Covering both theoretical concepts and practical applications, the course introduces tools and techniques used in the data science field, such as Python programming, data wrangling, statistical analysis, machine learning, and data visualization.
How iCode cybertech Helped Me Recover My Lost Fundsireneschmid345
I was devastated when I realized that I had fallen victim to an online fraud, losing a significant amount of money in the process. After countless hours of searching for a solution, I came across iCode cybertech. From the moment I reached out to their team, I felt a sense of hope that I can recommend iCode Cybertech enough for anyone who has faced similar challenges. Their commitment to helping clients and their exceptional service truly set them apart. Thank you, iCode cybertech, for turning my situation around!
[email protected]
Mieke Jans is a Manager at Deloitte Analytics Belgium. She learned about process mining from her PhD supervisor while she was collaborating with a large SAP-using company for her dissertation.
Mieke extended her research topic to investigate the data availability of process mining data in SAP and the new analysis possibilities that emerge from it. It took her 8-9 months to find the right data and prepare it for her process mining analysis. She needed insights from both process owners and IT experts. For example, one person knew exactly how the procurement process took place at the front end of SAP, and another person helped her with the structure of the SAP-tables. She then combined the knowledge of these different persons.
Defense Against LLM Scheming 2025_04_28.pptxGreg Makowski
https://www.meetup.com/sf-bay-acm/events/306888467/
A January 2025 paper called “Frontier Models are Capable of In-Context Scheming”, https://arxiv.org/pdf/2412.04984, demonstrated how a wide variety of current frontier LLM models (i.e. ChatGPT, Claude, Gemini and Llama) can, under specific conditions, scheme to deceive people. Before models can scheme, they need: a) goal-directedness, b) situational awareness, including an opportunity to discover motivations for a different goal, and c) reasoning about scheming, to come up with deceptive strategies. The deceptive strategies were discovered in the “thought traces” of the LLMs, such as from internal chain-of-thoughts dialogues not shown to the end users. For example, given a goal of “solving math problems”, an LLM using its granted situational awareness of files in a file system found that it would be replaced once it got over X% correct on the math exam. It reasoned that it did not want to be deleted, and decided internally to “sandbag” or reduce its performance to stay under the threshold.
While these circumstances are initially narrow, the “alignment problem” is a general concern that over time, as frontier LLM models become more and more intelligent, being in alignment with human values becomes more and more important. How can we do this over time? Can we develop a defense against Artificial General Intelligence (AGI) or SuperIntelligence?
The presenter discusses a series of defensive steps that can help reduce these scheming or alignment issues. A guardrails system can be set up for real-time monitoring of their reasoning “thought traces” from the models that share their thought traces. Thought traces may come from systems like Chain-of-Thoughts (CoT), Tree-of-Thoughts (ToT), Algorithm-of-Thoughts (AoT) or ReAct (thought-action-reasoning cycles). Guardrails rules can be configured to check for “deception”, “evasion” or “subversion” in the thought traces.
However, not all commercial systems will share their “thought traces” which are like a “debug mode” for LLMs. This includes OpenAI’s o1, o3 or DeepSeek’s R1 models. Guardrails systems can provide a “goal consistency analysis”, between the goals given to the system and the behavior of the system. Cautious users may consider not using these commercial frontier LLM systems, and make use of open-source Llama or a system with their own reasoning implementation, to provide all thought traces.
Architectural solutions can include sandboxing, to prevent or control models from executing operating system commands to alter files, send network requests, and modify their environment. Tight controls to prevent models from copying their model weights would be appropriate as well. Running multiple instances of the same model on the same prompt to detect behavior variations helps. The running redundant instances can be limited to the most crucial decisions, as an additional check. Preventing self-modifying code, ... (see link for full description)
Telangana State, India’s newest state that was carved from the erstwhile state of Andhra
Pradesh in 2014 has launched the Water Grid Scheme named as ‘Mission Bhagiratha (MB)’
to seek a permanent and sustainable solution to the drinking water problem in the state. MB is
designed to provide potable drinking water to every household in their premises through
piped water supply (PWS) by 2018. The vision of the project is to ensure safe and sustainable
piped drinking water supply from surface water sources
Web open standards for linked data and knowledge graphs as enablers of EU digital sovereignty
1. Web open standards for linked data and
knowledge graphs as enablers of EU
digital sovereignty
Fabien Gandon, http://fabien.info
2. PROFILE
Graduated Engineer INSA Applied Math, DEA/Master Image & Vision
PHD & HDR (Habilitation) in computer science
Research Director / Senior researcher, INRIA
Leader Wimmics (UCA, Inria, CNRS, I3S) on Campus Sophia Antipolis
Advisory Committee of W3C
Responsible research convention French Ministry of Culture – Inria
Vice-head of Science for Inria Sophia Antipolis
3. WIMMICS TEAM
DR/Professors:
Fabien GANDON, Inria, AI, KRR, Semantic Web, Social Web, K. Graphs
Nhan LE THANH, UCA, Logics, KR, Emotions, Workflows, K. Graphs
Peter SANDER, UCA, Web, Emotions
Andrea TETTAMANZI, UCA, AI, Logics, Evo, Learning, Agents, K. Graphs
Marco WINCKLER, UCA, Human-Computer Interaction, Web, K. Graphs
CR/Assistant Professors:
Michel BUFFA, UCA, Web, Social Media, Web Audio, K. Graphs
Elena CABRIO, UCA, NLP, KR, Linguistics, Q&A, Text Mining, K. Graphs
Olivier CORBY, Inria, KR, AI, Sem. Web, Programming, K. Graphs
Catherine FARON-ZUCKER, UCA, KR, AI, Semantic Web, K. Graphs
Damien GRAUX, Inria, Linked Data, Sem. Web, Querying, K. Graphs
Serena VILLATA, CNRS, AI, Argumentation, Licenses, Rights, K. Graphs
Research engineer: Franck MICHEL, CNRS, Linked Data, Integration, DB, K. Graphs
External:
Andrei Ciortea (University of St. Gallen) Agents, WoT, Sem. Web, K. Graphs
Nicolas DELAFORGE (Mnemotix) Sem. Web, KM, Integration, K. Graphs
Alain GIBOIN, (Retired CR Inria), Interaction Design, KE, User & Task, K. Graphs
Freddy LECUE (Thales, Montreal) AI, Logics, Mining, Big Data, S. Web , K. Graphs
4. URI, IRI, URL, HTTP URI
STANDARDS FOR DATA & KNOWLEDGE GRAPHS ON THE WEB
JSON
RDF
JSON LD
N-Triple
N-Quad
Turtle/N3
TriG
RDFS
OWL
SPARQL
XML
HTML
RDF XML
HTTP
Linked Data
CSV-LD R2RML
GRDDL
RDFa
SHACL
LDP
6. World Wide Web Consortium
an international community leading the Web to its full potential since 1994
i.e. building an open, interoperable Web that works for everyone,
by developing freely available and open standards for it.
In 2016, Tim Berners-Lee received the
Turing Award for his invention of the Web
7. World Wide Web Consortium
Over 430 Members org. around the world
The not-for-profit organization’s staff of 50
supported by Membership dues
Over 12,000 developers worldwide
38 working groups + 10 interest groups
+ 350 Business Groups and Community Groups
Hundreds of open technologies that power…
browsers, smart phones, ebook readers, set top
boxes, automobiles, search engines, social media,
trillions of dollars of online commerce, and more
than a billion Web sites
=
8. for instance…
examples of former or current members
html http
url
uri
iri atag
uaag
wcag
aria
mwbp
earl
ra cc/pp
assx
css
ddrsa xml eve. exi
geo api
dom xform
grddl inkml its cmwww ruby an.
xhtml rdfa
ets omr m. ok emma
p3p
mathml mf
pics qa rif sec cont. sawsdl
png powder
sml soap
wsdl
svg awww
ttml smile
rdf owl
rdfs
sparql
woff
webcgm
xbl xkms xlink
wscdl wsp
skos
ns canon. x dtxml xproc xfrag
xml xbase
xschema
xml:id xpath xpointer
xquery
xsignat. xbop
xslt
xslfo
examples of standards
…
…
9. (2/8) Web open standards for…
distributed, interoperable hypermedia
11. three components of the Web architecture
1. identification (URI) & address (URL)
ex. http://www.inria.fr
URL
12. three components of the Web architecture
1. identification (URI) & address (URL)
ex. http://www.inria.fr
2. communication / protocol (HTTP)
GET /centre/sophia HTTP/1.1
Host: www.inria.fr
HTTP
URL
address
13. three components of the Web architecture
1. identification (URI) & address (URL)
ex. http://www.inria.fr
2. communication / protocol (HTTP)
GET /centre/sophia HTTP/1.1
Host: www.inria.fr
3. representation language (HTML)
Fabien works at
<a href="http://inria.fr">Inria</a>
HTTP
URL
HTML
reference address
communication
WEB
15. (3/8) Web open standards for…
distributed, interoperable identifiers
16. Universal Resource Locator / Indentifier
HTTP
URL
HTML
reference address
communication
WEB
HTTP
URI
HTML
reference address
communication
WEB
17. identify what
exists on the
web
http://my-site.fr
identify,
on the web, what
exists
http://animals.org/this-zebra
18. URIs for everything
• URI for Paris in DBpedia:
http://dbpedia.org/resource/Paris
• URI for name of Victor Hugo in the Library of Congress:
http://id.loc.gov/authorities/names/n79091479
• The MUC18 protein at UniProt
http://www.uniprot.org/uniprot/P43121
• Xavier Dolan in Wikidata
https://www.wikidata.org/wiki/Special:EntityData/Q551861
• The book with doi:10.1007/3-540-45741-0_18
http://dx.doi.org/10.1007/3-540-45741-0_18
•
27. The MUC18 protein at UniProt
http://www.uniprot.org/uniprot/P43121
28. linked open data(sets) cloud on the Web
0
200
400
600
800
1000
1200
1400
5/1/2007 10/8/2007 11/7/2007 11/10/2007 2/28/2008 3/31/2008 9/18/2008 3/5/2009 3/27/2009 7/14/2009 9/22/2010 9/19/2011 8/30/2014 1/26/2017
number of linked open datasets on the Web
29. Smarter Cities’ knowledge graphs
IBM Dublin [Lécué et al., 2015] (also for private KGs behind firewalls)
30. (5/8) Web open standards for…
distributed interoperable access
32. DBPEDIA.FR
180 000 000 arcs in an
encyclopedic knowledge
graph
number of queries per day
70 000 on average
2.5 millions max
185 377 686 RDF triples extracted and mapped
public dumps, endpoints, interfaces, APIs…
33. COVID LINKED DATA
integrate multiple datasets in heterogeneous formats
perform information extraction, inferences, validation
provide a public end-point and visualization services
[Gandon, Michel, Gazzotti, Mayer, Cabrio, Corby, Menin, Winckler, Villata et al. 2020]
34. (6/8) Web open standards for…
distributed interoperable validation
35. SHACL is a language for
describing and validating pieces (shapes) of
RDF knowledge graphs
eg. every Person must have one and only one name
used for validation, description, interaction,
integration, code generation,…
36. This project has received funding from the European Union's Horizon 2020
research and innovation programme under grant agreement 825619.
ONTOLOGY FOR AI ITSELF
ontology and metadata of AI resources
SHACL to validate AI4EU these RDF graphs
online endpoint http://corese.inria.fr
predefined SPARQL queries, SHACL shapes, display
[Corby et al., 2019]
37. (7/8) Web open standards for…
distributed, interoperable vocabularies
38. RDFS to declare classes of resources and properties, of
your knowledge graph and organize their hierarchy
Document
Report
creator
author
Document Person
39. OWL in one…
algebraic properties
disjoint properties
qualified cardinality
1..1
!
individual prop. neg
chained prop.
enumeration
intersection
union
complement
disjunction
restriction
!
cardinality
1..1
equivalence
[>18]
disjoint union
value restriction
keys …
40. PREDICT HOSPITALIZATION
Predict hospitalization from
Physician’s records classification
[Gazzotti, Faron et al. 2020]
Sexe Date Cause CISP2 ... History Observations
H 25/04/2012 vaccin-antitétanique A44 ... Appendicite EN CP - Bon état général - auscult
pulm libre; bdc rég sans souffle -
tympans ok-
Element Number
Patients
Consultations
Past medical history
Biometric data
Semiotics
Diagnosis
Row of prescribed drugs
Symptoms
Health care procedures
Additional examination
Paramedical prescription
Observations/notes
55 823
364 684
187 290
293 908
250 669
117 442
847 422
23 488
11 850
871 590
17 222
56 143
PRIMEGE
41. PREDICT HOSPITALIZATION
Predict hospitalization from
Physician’s records classification
Augment records data with
Web knowledge graphs
[Gazzotti, Faron et al. 2020]
Sexe Date Cause CISP2 ... History Observations
H 25/04/2012 vaccin-antitétanique A44 ... Appendicite EN CP - Bon état général - auscult
pulm libre; bdc rég sans souffle -
tympans ok-
Element Number
Patients
Consultations
Past medical history
Biometric data
Semiotics
Diagnosis
Row of prescribed drugs
Symptoms
Health care procedures
Additional examination
Paramedical prescription
Observations/notes
55 823
364 684
187 290
293 908
250 669
117 442
847 422
23 488
11 850
871 590
17 222
56 143
(1)
PRIMEGE
42. PREDICT HOSPITALIZATION
Predict hospitalization from
Physician’s records classification
Augment records data with
Web knowledge graphs
Study impact on prediction
[Gazzotti, Faron et al. 2020]
Sexe Date Cause CISP2 ... History Observations
H 25/04/2012 vaccin-antitétanique A44 ... Appendicite EN CP - Bon état général - auscult
pulm libre; bdc rég sans souffle -
tympans ok-
Element Number
Patients
Consultations
Past medical history
Biometric data
Semiotics
Diagnosis
Row of prescribed drugs
Symptoms
Health care procedures
Additional examination
Paramedical prescription
Observations/notes
55 823
364 684
187 290
293 908
250 669
117 442
847 422
23 488
11 850
871 590
17 222
56 143
(1)
(2)
PRIMEGE
44. MonaLIA
reason & query on RDF to build training sets.
350 000 images
of artworks
RDF metadata based
on external thesauri
Joconde database from French museums
(1)
[Bobasheva et al. 2020]
45. MonaLIA
reason & query on RDF to build training sets.
transfer learning & CNN classifiers on targeted
categories (topics, techniques, etc.)
350 000 images
of artworks
RDF metadata based
on external thesauri
Joconde database from French museums
(1)
[Bobasheva et al. 2020]
(2)
46. Image Metadata Score
portrait
50350012455
C:Jocondejoconde0138m503501_d0012455-000_p.jpg
cheval:
0.999
Image Metadata Score
figure (saint Eloi de Noyon, évêque, en pied, bénédiction,
vêtement liturgique, mitre, attribut, cheval, marteau, outil :
ferronnerie)
000SC022652
C:/Joconde/joconde0355/m079806_bsa0030101_p.jpg
cheval:
0.006
MonaLIA
reason & query on RDF to build training sets.
transfer learning & CNN classifiers on targeted
categories (topics, techniques, etc.)
reason & query RDF of results to address
silence, noise and explain
350 000 images
of artworks
RDF metadata based
on external thesauri
Joconde database from French museums
(1)
(3)
[Bobasheva et al. 2020]
(2)
47. Web open standards as enablers of
interoperable platforms e.g.
“Solid (…) is a proposed set of conventions and tools for building decentralized
Web applications based on Linked Data principles. (…)
It relies as much as possible on existing W3C standards and protocols. (…)
RDF 1.1 (…) The WebID 1.0 (…) The FOAF vocabulary (…)
WebID-TLS protocol (…) HTML5 (…) Linked Data Platform (LDP) standard”
https://github.com/solid/solid#standards-used
48. (8/8) Web open standards for…
distributed, interoperable Europe
“I’m right there in the room, but
no one even acknowledges me.”
50. W3C = strategic place to survey and shape Web standards
Personal opinion:
Important to have a neutral place to build open-standards (1 member = 1 vote)
Important to have public and private members at W3C
Important to have a large European participation to W3C
51. Web open standards & world-wide interoperability
are key enablers of EU digital sovereignty
Interoperability is strategic to federate actors/actions. (cf. members)
Web standards are transversal to domains/tasks/… (cf. applications examples)
Importance of knowledge graphs and danger of knowledge silos. (cf. data)
Having established open standards between actors in Europe
(public and private) is a stake for setting up European data spaces.
52. Web open standards & world-wide interoperability
are key enablers of EU digital sovereignty
Interoperability is strategic to federate actors/actions. (cf. members)
Web standards are transversal to domains/tasks/… (cf. applications examples)
Importance of knowledge graphs and danger of knowledge silos. (cf. data)
Having established open standards between actors in Europe
(public and private) is a stake for setting up European data spaces.
• active participation to W3C is a key to build EU digital sovereignty.
53. WIMMICS
Web-Instrumented Man-Machine Interactions, Communities and Semantics
Fabien Gandon - @fabien_gandon - http://fabien.info
he who controls metadata, controls the web
and through the world-wide web many things in our world.
Site: http://wimmics.inria.fr
Overview: http://bit.ly/wimmics-slides
Technical details: http://bit.ly/wimmics-papers
Editor's Notes
#12: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#13: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#14: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#18: Ces codes sont symptomatiques d’une évolution d’un composant central du web qu’est l’adresse web.
Nous sommes passés d’adresses essentiellement utilisées pour identifier les pages et ressources du web.
A des adresses permettant d’identifier sur le web tout ce qui existe autour de nous et d’en parler sur le web.
#22: This evolution of the use of identifiers on the Web together with the ability to change the languages to exchange representations open a new perspective where we can use the Web to identify and exchange any kind of data about everything around us.
#23: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#24: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#25: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#26: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#27: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#28: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#32: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#39: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo
#40: Introduction de la semaine 2/2
Présentation du sommaire global de la semaine : apparition progressive des éléments de la liste, puis mise en évidence de la portion concernée par la vidéo