Sridhar Iyengar, IBM Distinguished Engineer at the IBM T. J. Watson Research Center, presention “Semantic PDF Processing & Document Representation” as part of the Cognitive Systems Institute Group Speaker Series.
Teaching cognitive computing with ibm watsondiannepatricia
Ralph Badinelli, Lenz Chair in the Department of Business Information Technology, Pamplin College of Business of Virginia Tech. presented "Teaching Cognitive Computing with IBM Watson" as part of the Cognitive Systems Institute Speaker Series.
Mridul Makhija has a B.Tech in Information Technology from Maharaja Institute of Technology. He currently works as a Machine Learning Engineer at CDAC Noida where he applies machine learning to predict patient volumes and blood bank requirements for AIIMS. Previously he worked as a Data Analyst at Bharti Airtel and held internships at Ericsson India and Cosco India. He has strong skills in Python, C++, data analysis, machine learning algorithms and deep learning. He has completed multiple personal projects applying machine learning and natural language processing. He has held leadership roles with Rotaract Club of Delhi and Interact Club and has participated in drama, volleyball and cricket competitions.
Tamanna Bhatt is a computer engineering graduate seeking a job where her work and ideas are appreciated. She has skills in Java, C, C++, C#.NET, Android development, HTML, CSS, JavaScript, jQuery, and MySQL. She earned a BE in computer engineering from Vadodara Institute of Engineering with a CGPA of 8.14. For her academic project, she developed a social networking Android app called Promarket that connects organizations and freelancers. She also prepared system requirements for an online government system project. She has visited BISAG and BSNL for exposure to geospatial and telecom systems.
Machine learning with an effective tools of data visualization for big dataKannanRamasamy25
Arthur Samuel (1959) :
"Field of study that gives computers the ability to learn without being explicitly programmed“
Tom Mitchell (1998) :
“A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E”.
There are several ways to implement machine learning algorithms
Automating automation
Getting computers to program themselves
Writing software is the bottleneck
Let the data do the work instead!
This document discusses artificial intelligence and its applications post-COVID 19. It is presented by Dr. Priti Srinivas Sajja from the department of computer science at Sardar Patel University. The document covers various topics related to AI such as its nature, symbolic AI, bio-inspired computing, applications in areas like healthcare, education, and examples of AI systems.
This document provides an introduction to data visualization. It discusses the importance of data visualization for clearly communicating complex ideas in reports and statements. The document outlines the data visualization process and different types of data and relationships that can be visualized, including quantitative and qualitative data. It also discusses various formats for visualizing data, with the goal of helping readers understand data visualization and how to create interactive visuals and analyze data.
Shivani Jain seeks a position as an IT professional to utilize her technical and intellectual abilities. She has a M.Tech in Information Technology from GGS Indraprastha University with 76.03% and a B.Tech in Information Technology from HMR Institute of Technology and Management with 74.2%. Her experience includes research work at ICAR-Indian Agricultural Statistical Research Institute and teaching at Mahan Institute of Technologies. She is proficient in languages like Java, C++, HTML, and technologies like CloudAnalyst and CloudSim.
Naman Singhal completed his B.Tech in Computer Science and Technology from IIIT Hyderabad with a CGPA of 8.5. He has worked on several projects related to power systems, document annotation, recommendation systems, search engines, and compilers. His work experience includes internships at Mentor Graphics and as a teaching assistant at IIIT Hyderabad. He has technical skills in programming languages like C, C++, Python and web technologies like HTML, CSS, PHP.
This Presentation has been presented in the Choice 2010 counseling event for IIT/NIT aspirants. In this presentation, Mr.K.RamaChandra Reddy ( CEO, MosChip Semiconductor Technology, India) explains the potential of Electronics Engineering and various career opportunities that are available for students.
Sakshi Sharma is a senior software developer with over two years of experience in HR, healthcare and retail industries. She has excellent troubleshooting skills and is able to analyze code to engineer well-researched, cost-effective solutions. She received a BE in Information Technology from Gyan Ganga Institute of Technology and Sciences. Currently she works as a lead developer at UST Global, handling Python scripting and creating workflows to meet requirements and deadlines. She has strong skills in Python, Django, Flask, MongoDB, machine learning and more.
Computational thinking (CT) is a problem-solving process that involves decomposition, pattern recognition, abstraction, and algorithm design. CT can be used to solve problems across many disciplines. The key principles of CT are: 1) Decomposition, which is breaking down complex problems into smaller parts; 2) Pattern recognition, which is observing patterns in data; 3) Abstraction, which identifies general principles; and 4) Algorithm design, which develops step-by-step instructions. CT is a concept that focuses on problem-solving techniques, while computer science is the application of those techniques through programming. CT can be applied to solve problems in any field, while computer science specifically implements computational solutions.
Visual Analytics for User Behaviour Analysis in Cyber SystemsCagatay Turkay
Slides for my short talk at the Alan Turing Institute at the "Visualisation for Data Science and AI" workshop (https://www.turing.ac.uk/events/visualization-data-science-and-ai).
The talk discusses a role for visualization to support decision making with algorithms and walks through an example of our EC H2020 funded DiSIEM research project.
CEN4722 HUMAN COMPUTER INTERACTIONS:
Please read Box 8.1: Use and abuse of numbers on Page 277 view the video on Data visualization. Will data visualization help us make better decisions? What are the downsides?
Entity-Relationship Extraction from Wikipedia Unstructured Text - OverviewRadityo Eko Prasojo
This is an overview presentation about my PhD research, not a very technical one. This was presented in the open session of WebST'16 Summer School in Web Science, July 2016, Bilbao - Spain.
Types of customer feedback, how easy are they to collect, analyse and how insightful are they?
Why analyzing customer feedback is important?
Why is it hard to analyze free-text customer feedback?
What approaches are there to make sense of customer feedback (manual coding, word clouds, text categorization, topic modeling, themes extraction) -- and what are their limitations?
Which AI methods can help with the challenges in customer feedback analysis.
Natural Language Processing (NLP) practitioners often have to deal with analyzing large corpora of unstructured documents and this is often a tedious process. Python tools like NLTK do not scale to large production data sets and cannot be plugged into a distributed scalable framework like Apache Spark or Apache Flink.
The Apache OpenNLP library is a popular machine learning based toolkit for processing unstructured text. Combining a permissive licence, a easy-to-use API and set of components which are highly customize and trainable to achieve a very high accuracy on a particular dataset. Built-in evaluation allows to measure and tune OpenNLP’s performance for the documents that need to be processed.
From sentence detection and tokenization to parsing and named entity finder, Apache OpenNLP has the tools to address all tasks in a natural language processing workflow. It applies Machine Learning algorithms such as Perceptron and Maxent, combined with tools such as word2vec to achieve state of the art results. In this talk, we’ll be seeing a demo of large scale Name Entity extraction and Text classification using the various Apache OpenNLP components wrapped into Apache Flink stream processing pipeline and as an Apache NiFI processor.
NLP practitioners will come away from this talk with a better understanding of how the various Apache OpenNLP components can help in processing large reams of unstructured data using a highly scalable and distributed framework like Apache Spark/Apache Flink/Apache NiFi.
Pipeline for automated structure-based classification in the ChEBI ontologyJanna Hastings
Presented at the ACS in Dallas: ChEBI is a database and ontology of chemical entities of biological interest, organised into a structure-based and role-based classification hierarchy. Each entry is extensively annotated with a name, definition and synonyms, other metadata such as cross-references, and chemical structure information where appropriate. In addition to the
classification hierarchy, the ontology also contains diverse chemical and ontological relationships. While ChEBI is primarily manually maintained, recent developments have focused on improvements in curation through partial automation of common tasks. We will describe a pipeline we have developed for structure-based classification of chemicals into the ChEBI structural classification. The pipeline connects class-level structural knowledge encoded in Web Ontology Language (OWL) axioms as an extension to the ontology, and structural information specified in standard MOLfiles. We make use of the Chemistry Development Kit, the OWL API and the OWLTools library. Harnessing the pipeline, we are able to suggest the best structural classes for the classification of novel structures within the ChEBI ontology.
Knowledge representation and reasoning (KR) is the field of artificial intelligence (AI) dedicated to representing information about the world in a form that a computer system can utilize to solve complex tasks such as diagnosing a medical condition or having a dialog in a natural language
Natural Language Processing and Graph Databases in LumifyCharlie Greenbacker
Lumify is an open source platform for big data analysis and visualization, designed to help organizations derive actionable insights from the large volumes of diverse data flowing through their enterprise. Utilizing both Hadoop and Storm, it ingests and integrates virtually any kind of data, from unstructured text documents and structured datasets, to images and video. Several open source analytic tools (including Tika, OpenNLP, CLAVIN, OpenCV, and ElasticSearch) are used to enrich the data, increase its discoverability, and automatically uncover hidden connections. All information is stored in a secure graph database implemented on top of Accumulo to support cell-level security of all data and metadata elements. A modern, browser-based user interface enables analysts to explore and manipulate their data, discovering subtle relationships and drawing critical new insights. In addition to full-text search, geospatial mapping, and multimedia processing, Lumify features a powerful graph visualization supporting sophisticated link analysis and complex knowledge representation.
Charlie Greenbacker, Director of Data Science at Altamira, will provide an overview of Lumify and discuss how natural language processing (NLP) tools are used to enrich the text content of ingested data and automatically discover connections with other bits of information. Joe Ferner, Senior Software Engineer at Altamira, will describe the creation of SecureGraph and how it supports authorizations, visibility strings, multivalued properties, and property metadata in a graph database.
1) The workshop discussed developing ontologies to represent mental functioning and disease, including modules for mental diseases, emotions, and related domains.
2) Ontologies provide standard vocabularies and computable definitions to facilitate data sharing and aggregation across studies and databases in areas like neuroscience, psychiatry, and genetics.
3) Relationships between ontology concepts can represent mechanisms and pathways involved in mental processes and diseases to enable new insights through automated reasoning.
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...semanticsconference
The NXTM Project is a research project between a university and IT company aimed at developing technology to analyze unstructured data streams and extract structured information. It involves processing documents through various analysis engines to identify semantics and link related data. The extracted structured data is stored in a database and made searchable through a semantic search engine. Search results are interactively represented as a graph to discover related information. The goal is to help small businesses extract valuable insights from unstructured data sources.
Ontology is the study of being or reality. It deals with questions about what entities exist and how they can be grouped, related within a hierarchy, and subdivided according to similarities and differences. There are differing philosophical views about the nature of reality, including whether reality is objective and exists independently of human observation, or is subjective and constructed through human experiences and social interactions. Ontological questions also concern whether social entities should be viewed as objective, external realities or as social constructions.
This document summarizes a workshop on data integration using ontologies. It discusses how data integration is challenging due to differences in schemas, semantics, measurements, units and labels across data sources. It proposes that ontologies can help with data integration by providing definitions for schemas and entities referred to in the data. Core challenges discussed include dealing with multiple synonyms for entities and relationships between biological entities that depend on context. The document advocates for shared community ontologies that can be extended and integrated to facilitate flexible and responsive data integration across multiple sources.
Artificial intelligence has the potential to significantly boost economic growth rates through its role as a capital-labor hybrid and its ability to accelerate innovation. AI can drive growth via three mechanisms: intelligent automation by adapting to automate complex tasks at scale, labor and capital augmentation by helping humans focus on higher value work and improving efficiency, and innovation diffusion by generating new ideas and revenue streams from data. For economies to fully benefit from AI, governments must prepare citizens and policy for integration with machine intelligence, encourage AI-driven regulation, advocate ethical guidelines for AI development, and address potential redistribution effects of job disruption.
1) Edge AI allows running AI models directly on devices to enable real-time decision making when latency is critical. Examples of edge AI applications include access control, quality control, and smart traffic lights.
2) Popular hardware for edge AI development includes Google's AIY Vision Kit and Amazon's DeepLens camera. Models are typically compressed using techniques like MobileNets to run efficiently on embedded devices.
3) The document demonstrates object detection models for tasks in the Audi Autonomous Driving Cup, achieving average inference times of around 11 milliseconds.
We develop custom Image Recognition systems for Aerospace and defence applications. Using algorithms like Deep Convolutional Neural Networks and Regional Convolutional Neural Networks.
Our algorithms for Target Recognition and Tracking are designed from the beginning to be run on embedded systems. We target both GPU and FPGA devices.
To Train and Validate our algorithms we developed a process to generate photorealistic 3D environments.
Those 3D Environments are used to produce realistic video streams of the targets in different environmental conditions (lighting, adverse meteorological conditions, camouflage, point-of-view).
The same technology can be used to Train and Test Automotive Vision Systems.
This presentation discusses developing custom image recognition systems for automotive and defense applications using deep learning techniques. It describes generating 3D environments to realistically simulate environmental and system conditions for training and validating vision algorithms. These virtual environments can be used to test automotive vision systems. The presentation also covers algorithm development targeting GPU and FPGA hardware, including a proposed integrated development environment for designing convolutional neural networks on FPGAs.
Naman Singhal completed his B.Tech in Computer Science and Technology from IIIT Hyderabad with a CGPA of 8.5. He has worked on several projects related to power systems, document annotation, recommendation systems, search engines, and compilers. His work experience includes internships at Mentor Graphics and as a teaching assistant at IIIT Hyderabad. He has technical skills in programming languages like C, C++, Python and web technologies like HTML, CSS, PHP.
This Presentation has been presented in the Choice 2010 counseling event for IIT/NIT aspirants. In this presentation, Mr.K.RamaChandra Reddy ( CEO, MosChip Semiconductor Technology, India) explains the potential of Electronics Engineering and various career opportunities that are available for students.
Sakshi Sharma is a senior software developer with over two years of experience in HR, healthcare and retail industries. She has excellent troubleshooting skills and is able to analyze code to engineer well-researched, cost-effective solutions. She received a BE in Information Technology from Gyan Ganga Institute of Technology and Sciences. Currently she works as a lead developer at UST Global, handling Python scripting and creating workflows to meet requirements and deadlines. She has strong skills in Python, Django, Flask, MongoDB, machine learning and more.
Computational thinking (CT) is a problem-solving process that involves decomposition, pattern recognition, abstraction, and algorithm design. CT can be used to solve problems across many disciplines. The key principles of CT are: 1) Decomposition, which is breaking down complex problems into smaller parts; 2) Pattern recognition, which is observing patterns in data; 3) Abstraction, which identifies general principles; and 4) Algorithm design, which develops step-by-step instructions. CT is a concept that focuses on problem-solving techniques, while computer science is the application of those techniques through programming. CT can be applied to solve problems in any field, while computer science specifically implements computational solutions.
Visual Analytics for User Behaviour Analysis in Cyber SystemsCagatay Turkay
Slides for my short talk at the Alan Turing Institute at the "Visualisation for Data Science and AI" workshop (https://www.turing.ac.uk/events/visualization-data-science-and-ai).
The talk discusses a role for visualization to support decision making with algorithms and walks through an example of our EC H2020 funded DiSIEM research project.
CEN4722 HUMAN COMPUTER INTERACTIONS:
Please read Box 8.1: Use and abuse of numbers on Page 277 view the video on Data visualization. Will data visualization help us make better decisions? What are the downsides?
Entity-Relationship Extraction from Wikipedia Unstructured Text - OverviewRadityo Eko Prasojo
This is an overview presentation about my PhD research, not a very technical one. This was presented in the open session of WebST'16 Summer School in Web Science, July 2016, Bilbao - Spain.
Types of customer feedback, how easy are they to collect, analyse and how insightful are they?
Why analyzing customer feedback is important?
Why is it hard to analyze free-text customer feedback?
What approaches are there to make sense of customer feedback (manual coding, word clouds, text categorization, topic modeling, themes extraction) -- and what are their limitations?
Which AI methods can help with the challenges in customer feedback analysis.
Natural Language Processing (NLP) practitioners often have to deal with analyzing large corpora of unstructured documents and this is often a tedious process. Python tools like NLTK do not scale to large production data sets and cannot be plugged into a distributed scalable framework like Apache Spark or Apache Flink.
The Apache OpenNLP library is a popular machine learning based toolkit for processing unstructured text. Combining a permissive licence, a easy-to-use API and set of components which are highly customize and trainable to achieve a very high accuracy on a particular dataset. Built-in evaluation allows to measure and tune OpenNLP’s performance for the documents that need to be processed.
From sentence detection and tokenization to parsing and named entity finder, Apache OpenNLP has the tools to address all tasks in a natural language processing workflow. It applies Machine Learning algorithms such as Perceptron and Maxent, combined with tools such as word2vec to achieve state of the art results. In this talk, we’ll be seeing a demo of large scale Name Entity extraction and Text classification using the various Apache OpenNLP components wrapped into Apache Flink stream processing pipeline and as an Apache NiFI processor.
NLP practitioners will come away from this talk with a better understanding of how the various Apache OpenNLP components can help in processing large reams of unstructured data using a highly scalable and distributed framework like Apache Spark/Apache Flink/Apache NiFi.
Pipeline for automated structure-based classification in the ChEBI ontologyJanna Hastings
Presented at the ACS in Dallas: ChEBI is a database and ontology of chemical entities of biological interest, organised into a structure-based and role-based classification hierarchy. Each entry is extensively annotated with a name, definition and synonyms, other metadata such as cross-references, and chemical structure information where appropriate. In addition to the
classification hierarchy, the ontology also contains diverse chemical and ontological relationships. While ChEBI is primarily manually maintained, recent developments have focused on improvements in curation through partial automation of common tasks. We will describe a pipeline we have developed for structure-based classification of chemicals into the ChEBI structural classification. The pipeline connects class-level structural knowledge encoded in Web Ontology Language (OWL) axioms as an extension to the ontology, and structural information specified in standard MOLfiles. We make use of the Chemistry Development Kit, the OWL API and the OWLTools library. Harnessing the pipeline, we are able to suggest the best structural classes for the classification of novel structures within the ChEBI ontology.
Knowledge representation and reasoning (KR) is the field of artificial intelligence (AI) dedicated to representing information about the world in a form that a computer system can utilize to solve complex tasks such as diagnosing a medical condition or having a dialog in a natural language
Natural Language Processing and Graph Databases in LumifyCharlie Greenbacker
Lumify is an open source platform for big data analysis and visualization, designed to help organizations derive actionable insights from the large volumes of diverse data flowing through their enterprise. Utilizing both Hadoop and Storm, it ingests and integrates virtually any kind of data, from unstructured text documents and structured datasets, to images and video. Several open source analytic tools (including Tika, OpenNLP, CLAVIN, OpenCV, and ElasticSearch) are used to enrich the data, increase its discoverability, and automatically uncover hidden connections. All information is stored in a secure graph database implemented on top of Accumulo to support cell-level security of all data and metadata elements. A modern, browser-based user interface enables analysts to explore and manipulate their data, discovering subtle relationships and drawing critical new insights. In addition to full-text search, geospatial mapping, and multimedia processing, Lumify features a powerful graph visualization supporting sophisticated link analysis and complex knowledge representation.
Charlie Greenbacker, Director of Data Science at Altamira, will provide an overview of Lumify and discuss how natural language processing (NLP) tools are used to enrich the text content of ingested data and automatically discover connections with other bits of information. Joe Ferner, Senior Software Engineer at Altamira, will describe the creation of SecureGraph and how it supports authorizations, visibility strings, multivalued properties, and property metadata in a graph database.
1) The workshop discussed developing ontologies to represent mental functioning and disease, including modules for mental diseases, emotions, and related domains.
2) Ontologies provide standard vocabularies and computable definitions to facilitate data sharing and aggregation across studies and databases in areas like neuroscience, psychiatry, and genetics.
3) Relationships between ontology concepts can represent mechanisms and pathways involved in mental processes and diseases to enable new insights through automated reasoning.
Adam Bartusiak and Jörg Lässig | Semantic Processing for the Conversion of Un...semanticsconference
The NXTM Project is a research project between a university and IT company aimed at developing technology to analyze unstructured data streams and extract structured information. It involves processing documents through various analysis engines to identify semantics and link related data. The extracted structured data is stored in a database and made searchable through a semantic search engine. Search results are interactively represented as a graph to discover related information. The goal is to help small businesses extract valuable insights from unstructured data sources.
Ontology is the study of being or reality. It deals with questions about what entities exist and how they can be grouped, related within a hierarchy, and subdivided according to similarities and differences. There are differing philosophical views about the nature of reality, including whether reality is objective and exists independently of human observation, or is subjective and constructed through human experiences and social interactions. Ontological questions also concern whether social entities should be viewed as objective, external realities or as social constructions.
This document summarizes a workshop on data integration using ontologies. It discusses how data integration is challenging due to differences in schemas, semantics, measurements, units and labels across data sources. It proposes that ontologies can help with data integration by providing definitions for schemas and entities referred to in the data. Core challenges discussed include dealing with multiple synonyms for entities and relationships between biological entities that depend on context. The document advocates for shared community ontologies that can be extended and integrated to facilitate flexible and responsive data integration across multiple sources.
Artificial intelligence has the potential to significantly boost economic growth rates through its role as a capital-labor hybrid and its ability to accelerate innovation. AI can drive growth via three mechanisms: intelligent automation by adapting to automate complex tasks at scale, labor and capital augmentation by helping humans focus on higher value work and improving efficiency, and innovation diffusion by generating new ideas and revenue streams from data. For economies to fully benefit from AI, governments must prepare citizens and policy for integration with machine intelligence, encourage AI-driven regulation, advocate ethical guidelines for AI development, and address potential redistribution effects of job disruption.
1) Edge AI allows running AI models directly on devices to enable real-time decision making when latency is critical. Examples of edge AI applications include access control, quality control, and smart traffic lights.
2) Popular hardware for edge AI development includes Google's AIY Vision Kit and Amazon's DeepLens camera. Models are typically compressed using techniques like MobileNets to run efficiently on embedded devices.
3) The document demonstrates object detection models for tasks in the Audi Autonomous Driving Cup, achieving average inference times of around 11 milliseconds.
We develop custom Image Recognition systems for Aerospace and defence applications. Using algorithms like Deep Convolutional Neural Networks and Regional Convolutional Neural Networks.
Our algorithms for Target Recognition and Tracking are designed from the beginning to be run on embedded systems. We target both GPU and FPGA devices.
To Train and Validate our algorithms we developed a process to generate photorealistic 3D environments.
Those 3D Environments are used to produce realistic video streams of the targets in different environmental conditions (lighting, adverse meteorological conditions, camouflage, point-of-view).
The same technology can be used to Train and Test Automotive Vision Systems.
This presentation discusses developing custom image recognition systems for automotive and defense applications using deep learning techniques. It describes generating 3D environments to realistically simulate environmental and system conditions for training and validating vision algorithms. These virtual environments can be used to test automotive vision systems. The presentation also covers algorithm development targeting GPU and FPGA hardware, including a proposed integrated development environment for designing convolutional neural networks on FPGAs.
Finely Chair talk: Every company is an AI company - and why Universities sho...Amit Sheth
Video: https://youtu.be/ZS8rGSzb_9I
The context of this talk is this statement from the host institution's provost: "We are trying to mobilize our campus activities around AI.” I connect academic initiatives in Interdisciplinary AI with industry needs.
--- Original abstract -----
Every company now is an AI company: Now, Near Future, or Distant Future?
Amit Sheth, AI Institute, University of South Carolina
“Every company now is an AI company. The industrial companies are changing, the supply chain…every single sector, it’s not only tech.” said Steven Pagliuca, CEO of Bain Capital at the 2019 World Economic Forum. With this statement as the context, I will provide an overview of AI landscape -- what AI capabilities are for real, what is being oversold, what is nonexistent, what is unlikely in our lifetime. I will also provide an anecdote-supported review through a broad variety of current and eminent applications of AI that rely on some of the well-developed and emerging AI capabilities. The objective is to help those considering AI applications start thinking of new business opportunities, new products and services, and new revenue/business models in the context of rapid penetration of AI technologies everywhere. I will seek to answer: Is AI just hype or something already happening? If it has not happened in your industry, is it impending? Do bad impacts of AI outweigh the good?
Alison Lowndes, Artificial Intelligence DevRel, Nvidia – Fueling the Artifici...Techsylvania
This document discusses how gaming and GPU computing can fuel advances in artificial intelligence. It provides examples of how AI is being used across many industries such as automotive, financial services, healthcare, and more. It also outlines how deep learning and neural networks are being used for applications like computer vision, natural language processing, robotics, and self-driving cars. Advances in GPU technology and deep learning frameworks are allowing AI to tackle increasingly complex challenges.
TechEvent 2019: Artificial Intelligence in Dev & Ops; Martin Luckow - TrivadisTrivadis
This document provides an overview of artificial intelligence trends and applications in development and operations. It discusses how AI is being used for rapid prototyping, intelligent programming assistants, automatic error handling and code refactoring, and strategic decision making. Examples are given of AI tools from Microsoft, Facebook, and Codota. The document also discusses challenges like interpretability of neural networks and outlines a vision of "Software 2.0" where programs are generated automatically to satisfy goals. It emphasizes that AI will transform software development over the next 10 years.
This document discusses intelligent image processing and its history, applications, and future directions. It begins with definitions of image processing and intelligence, then covers the history from artificial intelligence to new fields like neuromorphic engineering. Applications discussed include robotic vision, defense/security, medical imaging, manufacturing automation, and entertainment. The document also outlines the basic image processing chain and discusses current state-of-the-art examples like Google Goggles, Photosynth, and autonomous driving. It concludes by looking ahead to potential future areas like hyperspectral imaging, LIDAR, and biologically motivated processing.
This document provides an overview of artificial intelligence including definitions, issues, and applications. It defines AI as the study of intelligent agents that can perceive their environment and take actions to maximize success. Some key issues discussed are predictive recommendation systems and development of smarter objects like home assistants. Applications highlighted include IBM's Watson for health and education, Google Photos for image processing, Tesla's Autopilot, and MIT's Deepmoji for understanding emotions.
This document summarizes a presentation about the future of AI and Fabric for Deep Learning (FfDL). It discusses how deep learning has advanced due to increased data and computing power, but that commonsense reasoning will require more research. FfDL is introduced as an open source project that aims to make deep learning accessible and scalable across frameworks. It uses a microservices architecture on Kubernetes to manage training jobs efficiently. Research is ongoing to further develop explainable and robust AI capabilities.
This document summarizes Andre Freitas' talk on AI beyond deep learning. It discusses representing meaning from text at scale using knowledge graphs and embeddings. It also covers using neuro-symbolic models like graph networks on top of knowledge graphs to enable few-shot learning, explainability, and transportability. The document advocates that AI engineers should focus on representation design and evaluating multi-component NLP systems.
Rule-based Capture/Storage of Scientific Data from PDF Files and Export using...Stuart Chalk
Recently, the US government has mandated that publicly funded scientific research data be freely made available in a useable form, allowing integration of data in other systems. While this mandate has been articulated, existing publications and new papers (PDF) still do not provide accessible data, meaning that the usefulness is limited without human intervention.
This presentation outlines our efforts to extract scientific data from PDF files, using the PDFToText software and regular expressions (regex), and process it into a form that structures the data and its context (metadata). Extracted data is processed (cleaned, normalized), organized, and inserted into a contextually developed MySQL database. The data and metadata can then be output using a generic JSON-LD based scientific data model (SDM) under development in our laboratory.
Benefiting from Semantic AI along the data life cycleMartin Kaltenböck
Slides of 1 hour session of Martin Kaltenböck (CFO and Managing Partner of Semantic Web Company / PoolParty Software Ltd) on 19 March 2019 in Boston, US at the Enterprise Data World 2019, with its title: Benefiting from Semantic AI along the data life cycle.
Solutions for ADAS and AI data engineering using OpenPOWER/POWER systemsGanesan Narayanasamy
The ultimate goal of ADAS feature development is to make our roads safer and better suited for fully autonomous vehicles in the long run. Still, manufacturers and buyers shouldn’t underestimate the importance of ADAS for meeting current automotive challenges. The most significant impact of advanced driver assistance systems is in providing drivers with essential information and automating difficult and repetitive tasks. This increases safety for everyone on the road
As the adoption of AI technologies increases and matures, the focus will shift from exploration to time to market, productivity and integration with existing workflows. Governing Enterprise data, scaling AI model development, selecting a complete, collaborative hybrid platform and tools for rapid solution deployments are key focus areas for growing data scientist teams tasked to respond to business challenges. This talk will cover the challenges and innovations for AI at scale for the Enterprise focusing on the modernization of data analytics, the AI ladder and AI life cycle and infrastructure architecture considerations. We will conclude by viewing the benefits and innovation of running your modern AI and Data Analytics applications such as SAS Viya and SAP HANA on IBM Power Systems and IBM Storage in hybrid cloud environments.
Digital transformation with AI and process automation.
Prior consulting use cases in the domain of talent acquisition, e-commerce, e-Publishing and HR analytics.
Understanding Artificial Intelligence - Major concepts for enterprise applica...APPANION
Artificial Intelligence is a fundamental topic – for us as humans, as a society but also for businesses. For business executives and decision-makers, it is sometimes hard to keep up with rapidly evolving technologies as part of the day-to-day business. By providing this curated compilation of information about the fundamental aspects of AI, we want to captivate and inspire you to become more involved with the technology by better understanding the underlying concepts and value drivers of this technology
The document is an introduction to a series on document understanding presented by Mukesh Kala. It discusses what documents are, different types of documents including structured, semi-structured, and unstructured documents. It then covers topics like rule-based and model-based data extraction, optical character recognition, challenges in document understanding, and the document understanding framework which involves taxonomy, digitization, classification, extraction, validation, and training steps.
Cognitive systems institute talk 8 june 2017 - v.1.0diannepatricia
José Hernández-Orallo, Full Professor, Department of Information Systems and Computation at the Universitat Politecnica de València, presentation “Evaluating Cognitive Systems: Task-oriented or Ability-oriented?” as part of the Cognitive Systems Institute Speaker Series.
Building Compassionate Conversational Systemsdiannepatricia
Rama Akkiraju, Distinguished Engineer and Master Inventor at IBM, presention "Building Compassionate Conversational Systems" as part of the Cognitive Systems Institute Speaker Series.
“Artificial Intelligence, Cognitive Computing and Innovating in Practice”diannepatricia
Cristina Mele, Full Professor of Management at the University of Napoli “Federico II”, presentation as part of Cognitive Systems Institute Speaker Series
Eric Manser and Will Scott from IBM Research, presentation on "Cognitive Insights Drive Self-driving Accessibility" as part of the Cognitive Systems Institute Speaker Series
Roberto Sicconi and Malgorzata (Maggie) Stys, founders of TeleLingo, presented "AI in the Car" as part of the Cognitive Systems Institute Speaker Series.
Joining Industry and Students for Cognitive Solutions at Karlsruhe Services R...diannepatricia
Gerhard Satzger, Director of the Karlsruhe Service Research Institute and two former students and IBMers, Sebastian Hirschl and Kathrin Fitzer, presention"Joining Industry and Students for Cognitive Solutions at Karlsruhe Services Research Center" as part of the Cognitive Systems Institute Speaker Series.
170330 cognitive systems institute speaker series mark sherman - watson pr...diannepatricia
Dr. Mark Sherman, Director of the Cyber Security Foundations group at CERT within CMU’s Software Engineering Institute. , presention “Experiences Developing an IBM Watson Cognitive Processing Application to Support Q&A of Application Security Diagnostics” as part of the Cognitive Systems Institute Speaker Series.
“Fairness Cases as an Accelerant and Enabler for Cognitive Assistance Adoption”diannepatricia
Chuck Howell, Chief Engineer for Intelligence Programs and Integration at the MITRE Corporation, presentation “Fairness Cases as an Accelerant and Enabler for Cognitive Assistance Adoption” as part of the Cognitive Systems Institute Speaker Series.
From complex Systems to Networks: Discovering and Modeling the Correct Network"diannepatricia
This document discusses representing complex systems as higher-order networks (HON) to more accurately model dependencies. Conventionally, networks represent single entities at nodes, but HON breaks nodes into higher-order components carrying different relationship types. This captures dependencies beyond first order in a scalable way. The document presents applications of HON, including more accurately clustering global shipping patterns and ranking web pages based on clickstreams. HON provides a general framework for network analysis tasks like ranking, clustering and link prediction across domains involving complex trajectories, information flow, and disease spread.
Developing Cognitive Systems to Support Team Cognitiondiannepatricia
Steve Fiore from the University of Central Florida presented “Developing Cognitive Systems to Support Team Cognition” as part of the Cognitive Systems Institute Speaker Series
Kevin Sullivan from the University of Virginia presented: "Cyber-Social Learning Systems: Take-Aways from First Community Computing Consortium Workshop on Cyber-Social Learning Systems" as part of the Cognitive Systems Institute Speaker Series.
“IT Technology Trends in 2017… and Beyond”diannepatricia
William Chamberlin, IBM Distinguished Market Intelligence Professional, presented “IT Technology Trends in 2017… and Beyond” as part of the Cognitive Systems Institute Speaker Series on January 26, 2017.
Grady Booch proposes embodied cognition as placing Watson's cognitive capabilities into physical robots, avatars, spaces and objects. This would allow Watson to perceive the world through senses like vision and touch, and interact with it through movement and manipulation. The goal is to augment human abilities by giving Watson capabilities like seeing a patient's full medical condition or feeling the flow of a supply chain. Booch later outlines an "Self" architecture intended to power embodied cognitive systems with capabilities like learning, reasoning about others, and both involuntary and voluntary behaviors.
Kate is a machine intelligence platform that uses context aware learning to enable robots to walk farther in an unsupervised manner. Kate uses a biological architecture with a central pattern generator to coordinate actuation and contextual control to predict patterns and provide mitigation. In initial simulations, Kate was able to walk 8 times farther using context aware learning compared to without. Kate detects anomalies in its walking patterns and is able to mitigate issues to continue walking. This approach shows potential for using unsupervised learning from large correlated robot datasets to improve mobility.
1) Cognitive computing technologies can help address aging-related issues as over 65 populations increase in countries like Japan.
2) IBM Research has conducted extensive eldercare research including elderly vision simulation, accessibility studies, and conversation-based sensing to monitor health and provide family updates.
3) Future focus areas include using social, sensing and brain data with AI assistants to help the elderly live independently for longer through intelligent assistance, accessibility improvements, and early detection of cognitive decline.
The document discusses the development of cognitive assistants to help visually impaired people access real-world information and navigate the world. It describes technologies like localization, object recognition, mapping, and voice interaction that cognitive assistants can leverage. The goal is for assistants to augment human abilities by recognizing environments, objects, and providing contextual information. The document outlines a research project to develop such a cognitive navigation assistant and argues that accessibility needs have historically spurred innovations that become widely useful.
“Semantic Technologies for Smart Services” diannepatricia
Rudi Studer, Full Professor in Applied Informatics at the Karlsruhe Institute of Technology (KIT), Institute AIFB, presentation “Semantic Technologies for Smart Services” as part of the Cognitive Systems Institute Speaker Series, December 15, 2016.
Measuring Microsoft 365 Copilot and Gen AI SuccessNikki Chapple
Session | Measuring Microsoft 365 Copilot and Gen AI Success with Viva Insights and Purview
Presenter | Nikki Chapple 2 x MVP and Principal Cloud Architect at CloudWay
Event | European Collaboration Conference 2025
Format | In person Germany
Date | 28 May 2025
📊 Measuring Copilot and Gen AI Success with Viva Insights and Purview
Presented by Nikki Chapple – Microsoft 365 MVP & Principal Cloud Architect, CloudWay
How do you measure the success—and manage the risks—of Microsoft 365 Copilot and Generative AI (Gen AI)? In this ECS 2025 session, Microsoft MVP and Principal Cloud Architect Nikki Chapple explores how to go beyond basic usage metrics to gain full-spectrum visibility into AI adoption, business impact, user sentiment, and data security.
🎯 Key Topics Covered:
Microsoft 365 Copilot usage and adoption metrics
Viva Insights Copilot Analytics and Dashboard
Microsoft Purview Data Security Posture Management (DSPM) for AI
Measuring AI readiness, impact, and sentiment
Identifying and mitigating risks from third-party Gen AI tools
Shadow IT, oversharing, and compliance risks
Microsoft 365 Admin Center reports and Copilot Readiness
Power BI-based Copilot Business Impact Report (Preview)
📊 Why AI Measurement Matters: Without meaningful measurement, organizations risk operating in the dark—unable to prove ROI, identify friction points, or detect compliance violations. Nikki presents a unified framework combining quantitative metrics, qualitative insights, and risk monitoring to help organizations:
Prove ROI on AI investments
Drive responsible adoption
Protect sensitive data
Ensure compliance and governance
🔍 Tools and Reports Highlighted:
Microsoft 365 Admin Center: Copilot Overview, Usage, Readiness, Agents, Chat, and Adoption Score
Viva Insights Copilot Dashboard: Readiness, Adoption, Impact, Sentiment
Copilot Business Impact Report: Power BI integration for business outcome mapping
Microsoft Purview DSPM for AI: Discover and govern Copilot and third-party Gen AI usage
🔐 Security and Compliance Insights: Learn how to detect unsanctioned Gen AI tools like ChatGPT, Gemini, and Claude, track oversharing, and apply eDLP and Insider Risk Management (IRM) policies. Understand how to use Microsoft Purview—even without E5 Compliance—to monitor Copilot usage and protect sensitive data.
📈 Who Should Watch: This session is ideal for IT leaders, security professionals, compliance officers, and Microsoft 365 admins looking to:
Maximize the value of Microsoft Copilot
Build a secure, measurable AI strategy
Align AI usage with business goals and compliance requirements
🔗 Read the blog https://nikkichapple.com/measuring-copilot-gen-ai/
Master tester AI toolbox - Kari Kakkonen at Testaus ja AI 2025 ProfessioKari Kakkonen
My slides at Professio Testaus ja AI 2025 seminar in Espoo, Finland.
Deck in English, even though I talked in Finnish this time, in addition to chairing the event.
I discuss the different motivations for testing to use AI tools to help in testing, and give several examples in each categories, some open source, some commercial.
Protecting Your Sensitive Data with Microsoft Purview - IRMS 2025Nikki Chapple
Session | Protecting Your Sensitive Data with Microsoft Purview: Practical Information Protection and DLP Strategies
Presenter | Nikki Chapple (MVP| Principal Cloud Architect CloudWay) & Ryan John Murphy (Microsoft)
Event | IRMS Conference 2025
Format | Birmingham UK
Date | 18-20 May 2025
In this closing keynote session from the IRMS Conference 2025, Nikki Chapple and Ryan John Murphy deliver a compelling and practical guide to data protection, compliance, and information governance using Microsoft Purview. As organizations generate over 2 billion pieces of content daily in Microsoft 365, the need for robust data classification, sensitivity labeling, and Data Loss Prevention (DLP) has never been more urgent.
This session addresses the growing challenge of managing unstructured data, with 73% of sensitive content remaining undiscovered and unclassified. Using a mountaineering metaphor, the speakers introduce the “Secure by Default” blueprint—a four-phase maturity model designed to help organizations scale their data security journey with confidence, clarity, and control.
🔐 Key Topics and Microsoft 365 Security Features Covered:
Microsoft Purview Information Protection and DLP
Sensitivity labels, auto-labeling, and adaptive protection
Data discovery, classification, and content labeling
DLP for both labeled and unlabeled content
SharePoint Advanced Management for workspace governance
Microsoft 365 compliance center best practices
Real-world case study: reducing 42 sensitivity labels to 4 parent labels
Empowering users through training, change management, and adoption strategies
🧭 The Secure by Default Path – Microsoft Purview Maturity Model:
Foundational – Apply default sensitivity labels at content creation; train users to manage exceptions; implement DLP for labeled content.
Managed – Focus on crown jewel data; use client-side auto-labeling; apply DLP to unlabeled content; enable adaptive protection.
Optimized – Auto-label historical content; simulate and test policies; use advanced classifiers to identify sensitive data at scale.
Strategic – Conduct operational reviews; identify new labeling scenarios; implement workspace governance using SharePoint Advanced Management.
🎒 Top Takeaways for Information Management Professionals:
Start secure. Stay protected. Expand with purpose.
Simplify your sensitivity label taxonomy for better adoption.
Train your users—they are your first line of defense.
Don’t wait for perfection—start small and iterate fast.
Align your data protection strategy with business goals and regulatory requirements.
💡 Who Should Watch This Presentation?
This session is ideal for compliance officers, IT administrators, records managers, data protection officers (DPOs), security architects, and Microsoft 365 governance leads. Whether you're in the public sector, financial services, healthcare, or education.
🔗 Read the blog: https://nikkichapple.com/irms-conference-2025/
Content and eLearning Standards: Finding the Best Fit for Your-TrainingRustici Software
Tammy Rutherford, Managing Director of Rustici Software, walks through the pros and cons of different standards to better understand which standard is best for your content and chosen technologies.
Fully Open-Source Private Clouds: Freedom, Security, and ControlShapeBlue
In this presentation, Swen Brüseke introduced proIO's strategy for 100% open-source driven private clouds. proIO leverage the proven technologies of CloudStack and LINBIT, complemented by professional maintenance contracts, to provide you with a secure, flexible, and high-performance IT infrastructure. He highlighted the advantages of private clouds compared to public cloud offerings and explain why CloudStack is in many cases a superior solution to Proxmox.
--
The CloudStack European User Group 2025 took place on May 8th in Vienna, Austria. The event once again brought together open-source cloud professionals, contributors, developers, and users for a day of deep technical insights, knowledge sharing, and community connection.
"AI in the browser: predicting user actions in real time with TensorflowJS", ...Fwdays
With AI becoming increasingly present in our everyday lives, the latest advancements in the field now make it easier than ever to integrate it into our software projects. In this session, we’ll explore how machine learning models can be embedded directly into front-end applications. We'll walk through practical examples, including running basic models such as linear regression and random forest classifiers, all within the browser environment.
Once we grasp the fundamentals of running ML models on the client side, we’ll dive into real-world use cases for web applications—ranging from real-time data classification and interpolation to object tracking in the browser. We'll also introduce a novel approach: dynamically optimizing web applications by predicting user behavior in real time using a machine learning model. This opens the door to smarter, more adaptive user experiences and can significantly improve both performance and engagement.
In addition to the technical insights, we’ll also touch on best practices, potential challenges, and the tools that make browser-based machine learning development more accessible. Whether you're a developer looking to experiment with ML or someone aiming to bring more intelligence into your web apps, this session will offer practical takeaways and inspiration for your next project.
AI in Java - MCP in Action, Langchain4J-CDI, SmallRye-LLM, Spring AIBuhake Sindi
This is the presentation I gave with regards to AI in Java, and the work that I have been working on. I've showcased Model Context Protocol (MCP) in Java, creating server-side MCP server in Java. I've also introduced Langchain4J-CDI, previously known as SmallRye-LLM, a CDI managed too to inject AI services in enterprise Java applications. Also, honourable mention: Spring AI.
AI Emotional Actors: “When Machines Learn to Feel and Perform"AkashKumar809858
Welcome to the era of AI Emotional Actors.
The entertainment landscape is undergoing a seismic transformation. What started as motion capture and CGI enhancements has evolved into a full-blown revolution: synthetic beings not only perform but express, emote, and adapt in real time.
For reading further follow this link -
https://akash97.gumroad.com/l/meioex
Introducing FME Realize: A New Era of Spatial Computing and ARSafe Software
A new era for the FME Platform has arrived – and it’s taking data into the real world.
Meet FME Realize: marking a new chapter in how organizations connect digital information with the physical environment around them. With the addition of FME Realize, FME has evolved into an All-data, Any-AI Spatial Computing Platform.
FME Realize brings spatial computing, augmented reality (AR), and the full power of FME to mobile teams: making it easy to visualize, interact with, and update data right in the field. From infrastructure management to asset inspections, you can put any data into real-world context, instantly.
Join us to discover how spatial computing, powered by FME, enables digital twins, AI-driven insights, and real-time field interactions: all through an intuitive no-code experience.
In this one-hour webinar, you’ll:
-Explore what FME Realize includes and how it fits into the FME Platform
-Learn how to deliver real-time AR experiences, fast
-See how FME enables live, contextual interactions with enterprise data across systems
-See demos, including ones you can try yourself
-Get tutorials and downloadable resources to help you start right away
Whether you’re exploring spatial computing for the first time or looking to scale AR across your organization, this session will give you the tools and insights to get started with confidence.
UiPath Community Berlin: Studio Tips & Tricks and UiPath InsightsUiPathCommunity
Join the UiPath Community Berlin (Virtual) meetup on May 27 to discover handy Studio Tips & Tricks and get introduced to UiPath Insights. Learn how to boost your development workflow, improve efficiency, and gain visibility into your automation performance.
📕 Agenda:
- Welcome & Introductions
- UiPath Studio Tips & Tricks for Efficient Development
- Best Practices for Workflow Design
- Introduction to UiPath Insights
- Creating Dashboards & Tracking KPIs (Demo)
- Q&A and Open Discussion
Perfect for developers, analysts, and automation enthusiasts!
This session streamed live on May 27, 18:00 CET.
Check out all our upcoming UiPath Community sessions at:
👉 https://community.uipath.com/events/
Join our UiPath Community Berlin chapter:
👉 https://community.uipath.com/berlin/
DePIN = Real-World Infra + Blockchain
DePIN stands for Decentralized Physical Infrastructure Networks.
It connects physical devices to Web3 using token incentives.
How Does It Work?
Individuals contribute to infrastructure like:
Wireless networks (e.g., Helium)
Storage (e.g., Filecoin)
Sensors, compute, and energy
They earn tokens for their participation.
Adtran’s new Ensemble Cloudlet vRouter solution gives service providers a smarter way to replace aging edge routers. With virtual routing, cloud-hosted management and optional design services, the platform makes it easy to deliver high-performance Layer 3 services at lower cost. Discover how this turnkey, subscription-based solution accelerates deployment, supports hosted VNFs and helps boost enterprise ARPU.
European Accessibility Act & Integrated Accessibility TestingJulia Undeutsch
Emma Dawson will guide you through two important topics in this session.
Firstly, she will prepare you for the European Accessibility Act (EAA), which comes into effect on 28 June 2025, and show you how development teams can prepare for it.
In the second part of the webinar, Emma Dawson will explore with you various integrated testing methods and tools that will help you improve accessibility during the development cycle, such as Linters, Storybook, Playwright, just to name a few.
Focus: European Accessibility Act, Integrated Testing tools and methods (e.g. Linters, Storybook, Playwright)
Target audience: Everyone, Developers, Testers
From Legacy to Cloud-Native: A Guide to AWS Modernization.pptxMohammad Jomaa
“Semantic PDF Processing & Document Representation”
1. Future of Cognitive Computing and AI
Semantic PDF Processing and Knowledge
Representation
Sridhar Iyengar
Distinguished Engineer
Cognitive Computing Research
IBM T.J. Watson Research Center
[email protected]
9. Why is it hard? Variety of tables : 20-25 major table types
in discussion with just one major customer
Complex tables – graphical lines can
be misleading – is this 1, 2 or 3
tables ?
Table with
visual clues
only
Multi-row, multi-
column column
headers
Nested
row
headers
Tables with Textual
content
Table with
graphic
lines
Table
interleaved
with text and
charts
Complex multi-row,
multi-column column
headers identifiable
using graphical lines
and visual clues
10. Why is it hard? Variety in Image, Diagram Types
L. Lin et al. / Pattern Recognition 42 (2009) 1297--1307 1305
Fig. 8. ROC curves of the detection results for bicycle parts. Each graph shows the ROC curve of the results for a different part of the bicycle using just bottom-up information
and bottom-up + top-down information. We can see that the addition of top-down information greatly improves the results. We can also see that the bicycle wheel is the
most reliably detected object using only bottom-up cues, so we will look for that part first.
With a quick second glance, even the seat and handlebars may be
“seen”, though they are actually occluded. Our algorithm simulates
the top-down process (indicated by blue/green downward arrows in
Fig. 4) in a similar way, using the constructed And–Or graphs.
Verification of hypotheses: Each of the bottom-up proposals ac-
tivates a production rule that matches the terminal nodes in the
graph, and the algorithm predicts its neighboring nodes subject to
the learned relationships and node attributes. For example in Fig. 4,
a proposed circle will activate the rule that expands a wheel into
two rings. The algorithm then searches for another circle of propor-
tional radius, subject to the concentric relation with existing circle.
In Fig. 5(b), the wheels are already verified. The candidate frames
are then predicted with their ends affixed to the center points of the
wheels. Since we cannot tell the front wheels from the rear ones at
this moment, frames facing in two different directions are both pre-
dicted and put in the Open List. In Fig. 5(a), the triangle templates
are detected using a Generalized Hough Transform only when the
wheels are first verified and frames are predicted. If no neighboring
nodes are matched, the algorithm stops pursuing this proposal and
removes it from the Lists. Otherwise, if all of the neighboring nodes
are matched, the production rule is completed. The grouped nodes
are then put in the Closed List and lined up to be another bottom-up
proposal for the higher level. Note that we may have both bottom-
up and top-down information being passed about a particular pro-
posal as shown by the gray arrows in Fig. 3. In Fig. 4, the sub-parts
of the frame are predicted in the top-down phase from the frame
node (blue arrows); at the same time, they are also proposed in the
bottom-up phase based on the triangles we detected (red arrows).
Proposals with bidirectional supports such as these are more likely
to be accepted. After one particle is accepted from the Open List, any
other overlapping particles should update accordingly.
Template match: The pre-defined part templates, such as the bi-
cycle frames or teapot bodies, are represented by sub-sketch-graphs,
which are composed of a set of linked edgelets and junctions. Once a
template is proposed and placed at a location with initial attributes,
the template matching process is then activated. As shown in
10
PDF rendering
q .doc, .ppt rendering to .pdf keeps minimal structure formatting.
Geared towards visual fidelity
q Often .pdf is created by “screen scraping” or scanning or hybrid
ways that do not keep structure information.
Multi-modality: extremely rich information
q Images + Text + Tables both co-exist as well as form nested
hierarchies possibly with several levels
Nested table (numeric and
non-numeric + image)
Tabular representation
of images with pictorial
cross reference
Images + captions + cross references and
text that comments the image
11. Two major approaches to tackling PDF Processing
▪ Unsupervised Learning and out of the box PDF
processing
– Works well for a large class of domains with some compromise in
quality
▪ Supervised Learning with a graphical labelling tool
– Potential for improved quality when many similar documents are
available
Both approaches can be used together