2018 has been another breakthrough year for us as we traveled the world and launched another TWO World's First:
We would like to thank all of you for taking this sensational journey with us! We just can't wait to take on 2019!
Wishing you a Merry Christmas and a Fantastic 2019!
All of us at Babbobox
The term "Video Big Data" is rarely heard of. The reasons are pretty simple:
- It's difficult to extract data from videos
- It's difficult make sense of unstructured video data
Therefore, it is not an understatement to say that video is the most difficult medium to search and extract intelligence from. However, given the amount of videos are that generated daily in the public domain (e.g. YouTube) and private domain (e.g. broadcasters, CCTV, education, etc.), it is also not an understatement to say that video is the King of Content.
The objective of Big Data is to gain Business Intelligence. Video Big Data is no different. The obvious difference is the source and the type of data that can be extracted out from videos.
This Video Big Data Whitepaper aims to explain how we can extract value and intelligence from videos with a 3 step approach:
- Extract video data
- Transform unstructured video data
- Analyse to data into intelligence
With this whitepaper, we hope to share some of our knowledge and experiences working with Video Big Data. From our calculations, we estimate that Video Big Data will dwarf Big Data as we know it. Thus, the importance of this whitepaper. We hope you enjoy and benefit from it!
The VideoSpace Team
We are super excited about bringing our A.I. Video Search to Broadcast Asia after starting out in UK, US and China in 2018. It feels so good to be home!
Babbobox CEO, Alex Chan will be talking about "The Age of AI" and how it will transform the entire broadcast and media industry with Video Search, Personalized Content and Video Big Data.
We will also be making a big announcement and showcasing it during the show! We are pretty sure it will blow you away! So do drop by and say Hi!
The objective of Big Data is to gain Business Intelligence. Video Big Data is no different. The obvious difference is the source and the type of data that can be extracted out from videos. In there, lies the main challenges - Extraction, Transformation and Analysis.
In this instalment, we will explain why Artificial Intelligence is central to the “mess” in video big data.
In the first installment (Part 1), we explained:
- Why Video Big Data will absolutely dwarf current Big Data, and
- How Video is the most difficult medium to extract data
In the previous instalment (Part 2), we examined:
- the kind of data elements that we can extract from videos (speech, text, objects, activities, motion, faces, emotions)
But first, let’s examine why there is a mess in video data. The short explanation is because a large part of video data is unstructured data. In particular, data from speech and text. For example, text extracted from a 30 minutes news segment could cover multiple topics and events, mentioned numerous places and persons. To add to the complexities, we have to time-aligned when these words are spoken. In many ways, text (e.g. slide presentations that appear in videos) are the same.
Thus, we have to answer 2 key question:
- How do we meet sense of ‘messy’ video data?
- How can we extract knowledge or intelligence from that mess?
The answer lies in another form of Artificial Intelligence (A.I.) - the study of Natural Language Processing (NLP). That is because it can process and attempt to make sense of unstructured text in the following areas:
- Topic detection
- Key phrase extraction
- Sentiment analysis
The reason is because NLP can be used to turn unstructured video data into structured data. Only then can we start making sense and manipulating the data into either intelligence or actionable items like alerts, triggers, etc.
The field of Video Big Data is just starting. Without the advancement in multiple areas of Artificial Intelligence in multiple areas (Speech Recognition, Computer Vision, Facial Analysis, Text Analytics, etc.), Video Big Data wouldn’t even exist as it needs these fields to work in tandem or in sequence.
Given the rate that we are producing videos, alongside our ability to extract video data using A.I. The only way is up and we are not even close to uncovering the tip of Video Big Data iceberg.
Video Big Data will be bigger than BIG.
VideoSpace will be right in the middle of it all. Let’s put this prediction into a time capsule and revisit it in a few years.
In the last installment, we explained:
- Why Video Big Data will absolutely dwarf current Big Data
- How Video is the most difficult medium to extract data from
Which explains why Video Big Data remains a largely unexplored field. But also means the intense opportunities available because we have not even scrap the tip of this huge data iceberg.
In this installment, we will examine the kind of data elements that we can extract from videos.
In a hour of video, a person can say up to 9,000 words. So imagine the amount of data just from speech alone. However, the process of transcribing speech is filled with problems and we are currently only starting to get an acceptable level of accuracy.
Besides speech, text is probably the second most important element inside videos. For example, in a presentation or lecture, besides speech the speaker would augment the session with a set of slides. Or news tickers appearing during a news broadcast.
There are thousands of objects inside a video within different timeframe. Therefore, it can be quite challenging to identity what objects are in the video content and in which scene they appear in.
The difference between video and still images is motion. Different video scenes contain complex activities, such as “running in a group” or “driving a car”. Ability to extract activities will give a lot of insight what the videos are about. This includes offensive content that might contain nudity and profanity.
Detecting motion enables you to efficiently identify sections of interest within an otherwise long and uneventful video. That might sound simple, but what if you have 10,000 hours of videos to review every night? That’s a near impossible task to eyeball every video minute.
Detecting faces from videos adds face detection ability to any survelliance or CCTV system. This will be useful to analyze human traffic within a mall, street or even a restaurant or café. When we include facial recognition, it opens up another data dimension.
Emotion detection is an extension of the Face Detection that returns analysis on multiple emotional attributes from the faces detected. With emotion detection, one can gauge audience emotional response over a period of time.
This list of video data is certainly not exhaustive but is a definitely a good starting point to the field of Video Big Data. In the next installment, we will examine some of the techniques used to extract these video data.
The Babbobox Team
YouTube sees more than 300 hours of videos uploaded every minute. That's 432,000 hours in 1 day or 158 million hours in 1 year. That's 18,000 years worth of videos in a year. And that's just YouTube ONLY! If we add all other videos in the public domain, we wouldn't even know where to start with the numbers.
However, the even bigger numbers are actually hidden in the private domain from sources like broadcasters, surveillance cameras, GoPros, bodycams, smart devices, etc. We are recording videos at an unprecedented pace and scale.
There is one word to describe this phenomenon - BIG!
Which brings us to Video Big Data. Or should I say the lack of it. Even the term "Video Big Data" is rarely heard of. This stems from the inability to extract video data and making sense of it. But there is so much information embedded inside videos that is waiting to be discovered.
So the real question is... how can we extract value from videos?
However, the problem with video is that it is the most difficult medium to work with. There are a few reasons why:
- It is very difficult to extract various elements (speech, objects, faces, etc.) of video data.
- Each video element requires a different data extraction technique.
- It is very difficult to make sense of video data because of its unstructured nature
But there is hope yet. We will examine how we can tackle these problems and extract value from video big data in the next article.
Our launch - "World's First Video Search Engine with Interactive Results" in Birmingham (UK) was picked up by The Record and given some airtime. It feels great to be picked up and be given that bit of recognition for doing what we do to a global audience.
Click HERE for the article.
Note: The Record is a global magazine featuring the Best of Enterprise Technology on The Microsoft Platform.
Finally, 2 intensive days of MS Tech Summit in Birmingham... done and dusted. Absolutely the right decision to come to UK to do this. Massive event! Exactly the right platform to showcase our Video Search technologies.
Caught up with Scott Guthrie. Held so many in-depth discussion with so many UK enterprises, universities, government agencies, etc. If we have our way, our stuff might even end up in Scotland Yard! So let's see...
Good-bye Birmingham... Next stop, Trump-capital Washington in March! I'm excited already...
Birmingham, 24 January 2018: - Babbobox and Infini Videos officially announce the launch of the world’s first “Video Search Engine with Interactive Results” at the Microsoft Tech Summit held in Birmingham, United Kingdom today.
Both tech start-ups Babbobox and Infini Videos believe the future of video search lies in immediate content relevance. Video has proven to be the hardest medium to index because there is so much detail. Aside from the metadata that an editor may have typed in, most archived videos are essentially unstructured data. Often, this is because transcripts are not made and scripts are lost, or there isn’t sufficient timing information to align with the video.
To make sense of this data, techniques such as Speech Recognition, Video OCR, Image analysis and various Cognitive and Artificial Intelligence are applied to extract data from media. Since much of the videos in archives contains speech, therefore automatic transcription is a great first step in extracting data from media. With the transcript, an editor is able to search for timecodes in source videos, scrub through those sources, and manually locate viable scenes. This manual process is time-consuming, and not suitable for public use since a text search result does not make for a watchable video.
The innovative “Video Search Engine with Interactive Results” that Babbobox and Infini Videos have co-produced, allows a user to search a topic, and immediately view the search results as an interactive video. One is able to immediately choose scenes within the video that are relevant to the search. With the full-automation of the indexing and the output scene selection, productivity is enhanced and content for the public will scale up. The platform promises increased productivity as it will provide users with fine control over topics and sources, and allow editors to focus on the direction of the stories.
Babbobox and Infini Videos believe that this will be the future of Video Search.
About Babbobox (website: www.babbobox.com)
Babbobox launched two World’s First - World’s First True Unified Search Engine where it has the ability to index and “Search Everything” (all formats including video, audio and documents). Thus, positioning Babbobox to become "The Next Generation of Intelligent Storage". With VideoSpace, we created the World’s Frist Video-Search-as-a-Service. Thus, forming the foundations to enable a new breed of video services for the world.
About Infini Videos (website: www.infinivideos.com)
Infini Videos is a B2B online technology platform for the creation and delivery of HTML5 interactive videos. Infini Videos makes it easy to create engaging interactive videos as well as to access the rich data analytics offered on the platform. The company currently offers branching aka “Choose-your-own-adventure” and 360-degree types of interactivity. In addition to the technology platform, the company also provides specialized creative services as a one-stop solution for clients. Infini Videos is part of the Mediacorp’s MediaPreneur Incubator programme.
To find out more, please CONTACT US