Natural Language Processing combats manual text analysis
We’re constantly collecting more data, for example from camera images and text documents. This can provide us with relevant information. However, data is not always stored in a structured manner. This makes it difficult to retrieve the relevant information. Natural Language Processing (NLP) is an AI technique that tackles this problem.
What is natural language processing?
NLP combines the techniques of statistics with machine learning. This makes it possible to extract keywords from a text. We can then use this to make important classifications. TNO uses NLP to extract information from extensive, unstructured textual data in a more automated way.
TNO automatically creates taxonomies with natural language processing
You can use jargon to better streamline and standardise processes, for example in the form of a taxonomy or ontology. However, matching jargon within a field is a time-consuming exercise.
TNO uses NLP to identify important terms from a set of documents and determine their mutual relationships. We do this by:
- combining syntactic information (sentence construction)
- keyword extraction
- web sources
- semantic embedding methods
The taxonomy can then be used as input for an expert session.
Natural language processing is relevant for trend prediction
At TNO, we use our tools to automatically extract information from documents. We can also make predictions, such as in the foresight domain. Using the Horizon Scanner, we explore and extract from relevant websites, blogs and documents. This allows us to retrieve relevant information and to show trends.
Trend analysis shows us that the term deep learning is now being mentioned much more frequently within the computer vision domain than it was ten years ago. In addition, we can classify the documents automatically. For example, by a particular topic or field. We can also use blogs to conduct sentiment analysis and find out whether terms are being described more positively or negatively.
Timon BrussaardFunctie:Sr. Business Development Manager
Timon Brussaard is the Business Director at Leibniz Institute and Senior Business Development Manager at TNO. With a background in economics and change management, for the past 18 years, Timon has worked on business development, research, and consultancy processes in the rapidly changing ICT market. His expertise includes, norm engineering, information provision, AI, leadership, project management, communication, policy skills and marketing.
Christopher BrewsterFunctie:Senior scientist
Christopher Brewster is a Senior Scientist in the Data Science group and Professor of the Application of Emerging Tecnologies in the Institute of Data Science, Maastricht University. His research has focussed on the application of Semantic Technologies, Open and Linked Data, interoperability architectures and Data Governance, mostly to the food and agriculture domains.
Daniël WormFunctie:Senior consultant
Joris SijsFunction not known
Judith DijkFunctie:senior research scientist
Judith is specialised in extracting information from camera images. She now applies the subject of her PhD thesis in Physics, which she obtained 18 years ago, to her work as a research scientist at TNO, including in a research programme on camera systems for the Dutch Ministry of Defence.
Looking for another expert?View all experts
Responsible decision-making between people and machines
Bias in facial recognition and accidents with self-driving cars. AI must be developed further. The fastest way to do this is in close cooperation with people.
Knowledge representation and reasoning
Correct and unambiguous information is needed when making a decision. That is why we use AI technology called "knowledge representation & reasoning".
Robotics and autonomous agents
Robotics brings a future-proof industry a big step closer. For example, we are working on automatic path planning with AI techniques.
Fair machine learning
Fair machine learning is relevant to all kinds of discrimination and bias arising from the use of biased data. Read more!
Through Deep Vision, we’re developing AI algorithms to make automatic image analysis possible. Learn more!