In perhaps the ultimate cool endorsement, last week at WWDC Apple announced optical character recognition will be embedded in iOS 15. Named LiveText OCR, Apple describes it as “secure on-device intelligence to help you discover more in your photos, quickly find what you’re looking for, and relive special moments.”
We’ve been talking about multimedia data capture in ECM and BPM for at least 20 years. Recent advancements in computing power and deep learning, leveraged by innovative companies like AnyClip and Veritone, are finally moving into the information management mainstream. We expect to see this take off over the next 2-3 years.
Under the hood, Vantage utilizes Convolutional Neural Networks (CNN) that are pre-trained on hundreds of thousands of documents to extract visual features from a document. A Recurrent Neural Network (RNN) is used to extract semantic features of the text. While undoubtedly welcome to the large community of ABBYY users and the market at large, neither method is ground-breaking or particularly innovative.