Ten years ago, researchers thought that getting a computer to tell the difference between a cat and a dog would be almost impossible. Today, computer vision systems do it with greater than 99 percent accuracy. How? Joseph Redmon works on the YOLO (You Only Look Once) system, an open-source method of object detection that can identify objects in images and video — from zebras to stop signs — with lightning-quick speed. In a remarkable live demo, Redmon shows off this important step forward for applications like self-driving cars, robotics and even cancer detection.
Check out more TED talks:
The TED Talks channel features the best talks and performances from the TED Conference, where the world’s leading thinkers and doers give the talk of their lives in 18 minutes (or less). Look for talks on Technology, Entertainment and Design — plus science, business, global issues, the arts and more.
Follow TED on Twitter:
Like TED on Facebook:
Subscribe to our channel:
Ya çeviren arkadaşlar sağ olsunlar emek harcamışlar da her videoda mı hata olur çeviride
@4:25 You might have trained computers to act more like humans, but you can't train humans to act more like humans. XD
I thought that was a robot walking towards him at 6:45 😂 Lol
Hii brothers and sisters
It is usefull for freshers
Basic to advance
https://youtu.be/-RPl2rAKXc8
Very helpful video!
it should detect it as toy animals, its not real animals.
5:53 huge dining table. 😂
A skateboard at 3:22
who else saw skateboard at 3:22
Amazing!
I would like to know the specs of that laptop
See YOLO deployed to GTA V, the video is on my page 🙂
need this technology for my suit
This is so cool
This man is smart
7:10
It already knows more than we do. The stop sign is not a stop sign, it is a frisbee
Guys please check out his resumes. It is legendary
Quite amazing…
It saw it detect a stop sign as a frisbee.
Better not put that in a car.
Want to know more
What's the second detection algorithm that he executed? Is that Fast RCNN?
Does anyone know how it was ran on a phone ?
Image annotation interface(for object detection) check https://github.com/emadboctorx/labelpix
hello. how this system works ?
still doesn't solve the unsupervised learning problem because if you solved unsupervised learning, you would not need to pretrain any object at all, instead the program could just figure out on its own how to classify what object based on information alone from the outside having nothing to build on to begin with other than the method of the learning algorithms itself. example, what if i wanted to make the network recognize sound or words at the same time as detecting people and animals. what if i wanted to make the program think that the sound it heard was directly related to a portion of the screen and treat that portion as a separate object until the program had done this same process over and over and over making different version of part of the screen as that particular object until it can tell background from actual object. it means the program must find part of a image to be more of one object that other parts where each versions of this audio based guess of graphics narrow down what a object is as a separate object discarding what is not a object or another object from what its previous version was thinking was the object of iterest. let say a program just make a section of the image a cut based on aproximity and angle and start learning what ever pattern in it and decides its a dog even it could be part of a table or bed at the same time but over time of repetition is able to separate the real dog from other objects. what i means is what if you just start by training in junk with certain keywords of jumble and try to make the network over time detect real objects and classify real names to them by guessing. this way, you chould be able to make unsupervised learning work. just think of how a child think the name dog is spelled dol and over time learning its spelled dog. anyway if such a system like yolo could have incorporated somthing like i suggest, it could have a real advantage of realtime tracking of what at first will be detected junk and classified jumble. this way, you don''t need to pretrain any data but let it train itself over time or teach itself over time. what it would do would be like a human getting everything wrong at first past the baby stage and then become more and more human as it learns. i hope there is some real superintelligent human out there that want to make a unsupervised version of yolo that stated out as a complete idiot of a program and over time become very acurate in predicting what object is what, and with what name.
This is really great, but i had some installation issues while installing darkflow in my laptop.
Any one who know this willing to help me?
جيد جدا
can someone tell When I train YOLO for custom detection via my own dataset, it is again and again giving error that "STB Reason : Could not fopen" to the image path stored on Training line being exectuted
Now you can do this in 3 lines of code. Jeez, how far we come in 3 years
where can I get the maven dependency please if useable for java…