Readit News logoReadit News
bawana · 6 years ago
We are all essentially evil because 1) tomorrow will find some behavior that is accepted today as bad and we are all doing it 2) creating an AI that can manipulate people is too tempting for the predator in us to avoid (this is a REAL turing test - making an AI that can make tools out of humans and everything else.)

The only way to stay ahead of the 'evil' corrupting influence of new tech is to prevent it from widespread use controlled by a single entity. So, yolo is ok as long as you cannot deploy it in a cloud at scale.

So, just as nuclear weapons (the massive concentration of energy release at tremendously fast rates) are bad, so is a super AI/AGI (the massive computational ability at nanosecond scale).

No evil was ever perpetrated by institutions of learning - only business entities and governments who scaled up those discoveries caused evil.

And now for the flame bait- So, by this argument we should elect Luddites to govern us , especially ones that are not imaginative or creative.

haditab · 6 years ago
I believe this is exactly why pjreddie quit computer vision research. It must kill him to see such projects based off of his work.
tda · 6 years ago
Maybe someone else is also interested in some backgrounds on this https://news.ycombinator.com/item?id=22390093
Abishek_Muthian · 6 years ago
Thank you.
woah · 6 years ago
This detects any face, it does not identify people. It’s for stuff like autofocus, etc.
voqv · 6 years ago
Detecting where is a face in a picture is the first step that's necessary before detecting whose face is that.
vtange · 6 years ago
Autofocus can be for anything - Phone cameras, Surveillance cameras, Drone missile targeting systems, etc.
sudosysgen · 6 years ago
It's not really for autofocus. For autofocus, you need a model that can detect blurred images. You also need to operate on 36-42 bit data.

Furthermore, autofocus has already progressed from face detection to eye detection.

taneq · 6 years ago
Autofocus, autotarget, aimbots...
jonex · 6 years ago
Could you elaborate? What is the problem with the linked project? Training a slightly faster, smaller and less accurate version of an existing model?
chvid · 6 years ago
https://pjreddie.com/darknet/yolo/

Is that pjreddie used horses, dogs, and bicycles as training data? Not realising that his technology could also be used on human faces?

naasking · 6 years ago
I think that's the wrong approach. You don't stop the arms race by not investing in arms, you invest in countermeasures instead.

Deleted Comment

Deleted Comment

vffhfhf · 6 years ago
I also believe it was idiotic.

Focusing too much on negative aspect of thing will lead you nowhere.

Manhattan project gave >500 research papers. Gave us Iodine 131 and other radio nucleotide which we use in medicine. And gave us lasting peace. So was it bad or not?

Is gene editing bad? was the internet bad? Was the dude who invented round wheel bad?

aiilns · 6 years ago
Yes, using atomic bombs was bad. As others say it is debatable if it helped end the (already won?) war. And in any case, _lasting_ _piece_ where? Are middle east and Africa not part of this world?

I find your comment is entirely missing the point and low effort. Asking whether random techniques, inventions, inventors were "bad" or not makes us much sense as asking:

Is the sun bad? Were dinosaurs bad? Are the aliens bad? Is life bad?

pugworthy · 6 years ago
| And gave us lasting peace

I think the jury is still out on that. Or rather the trial is still underway.

cheesecracker · 6 years ago
You know who else can recognize faces? Humans! I guess we should cancel humans because of ethical concerns.
chii · 6 years ago
one human brain can only perform some 10-12 hrs of facial recognition work before needing to take a break, and also do so relatively low speeds.

One computer brain can be copied for free, and deployed to thousands of computer clusters and work 24/7 on facial recognition, at a fraction of the cost.

oh_sigh · 6 years ago
Because small models of face detection software would be released on github?
Tempest1981 · 6 years ago
I'm looking for something to run on a Raspberry Pi, to detect humans on a security camera. The built-in camera software has false triggering, esp. on windy days.

When looking at these projects, how do I figure out what hardware they're aimed at? This one mentions NVidia/CUDA.

Is there any sort of hardware abstraction layer that YOLO or R-CNNs can operate on? Can I use any of this code (or models) for my R-Pi?

dmm · 6 years ago
Chezck out the Frigate project. It uses the Coral tpu accelerator which could be used with the rpi.

https://github.com/blakeblackshear/frigate

earleybird · 6 years ago
'round about 2004 I built a pentium based 'motion' recorder. It kept a circular buffer of images that were spooled to the output stream when motion was detected. Motion was determined by optical flow iirc - the OpenCV call returned an array of blob center points, size, and velocity vector. If the blob was large enough and the velocity vector made sense (eg, horizontal as in walking or driving at an appropriate magnitude) it was considered motion. Reduced leaf flutter, branch waving false positives to effectively zero. No ML required. ML is too liberally applied without understanding how or why it has triggered. I've lost track of the original quote but the spirit of it is: "It's artificial intelligence while we don't understand it. Once we understand it, it's computer science"
Tempest1981 · 6 years ago
Interesting, I may try that out of curiosity. Although seeing Mask R-CNN demos is pretty intriguing, albeit expensive.

I remember a professor saying, "The definition of AI is: something that doesn't work"

w_t_payne · 6 years ago
You might want to investigate the Movidius Neural Compute stick (which you can use with a RPi), or the Nvidia Jetson Nano, which has a lot more oomph.
8fingerlouie · 6 years ago
I use the Movidius NCS and it's pretty much the optimal solution for light OpenCV work. It draws very little power, but as the Pi itself is powered by USB, don't expect to run much else off of USB on it.

I use it in the exact same scenario, a Raspberry Pi (Zero W) with a camera with motion detection and notifications on movement, my implementation may be specific though.

Each of my Raspberry Pi cameras runs motion (https://motion-project.github.io/index.html), and recorded files are stored on a NFS share. Each camera has it's own directory within this share (or rather each camera has it's own share within a parent directory), and the server then runs a python script that monitors for changed/added files, and runs object detection on the newly created/changed files.

If a person is detected in the file, it then proceeds to create a "screenshot" of the frame with the most/largest bounding box, and sends a notification through Pushover.net including the screenshot with bounding box.

There implementation is not quite as simple as described here, i.e. i use a "notification service" listening on MQTT for sending pushover notifications, but the gist of it is described above.

Edit: I should probably clarify that my cameras are based on Raspberry Pi Zero W. They have enough power to run motion at 720p - at around 30fps. Not great, but good enough for most applications. I've since migrated most to Unifi Protect instead. A little higher hardware cost, a lot better quality :)

mycall · 6 years ago
I found Sipeed MaixCube [1] to be pretty good at YOLO, lots of tech for only $25.

[1] https://www.seeedstudio.com/Sipeed-Maix-Cube-p-4553.html

rsaxvc · 6 years ago
You can use darknet with NNPACK on the RPis
DSingularity · 6 years ago
I love how every new YOLO project inevitably leads to the discussion of the ethics. At the very least more people will be wondering if they should also be taking ethics into consideration wrt their lines of work.

Pjreddie is a giant for this. It is a real contribution.

rgrieselhuber · 6 years ago
This also puts social distancing into perspective.
rocauc · 6 years ago
Purpose-built, small, and fast models appears to be the inevitable evolution for computer vision.

Where can the "Easy Set, "Medium Set, and "Hard Set" evaluations referenced in the "Wider Face Val" be found?

qiuqiu-dog · 6 years ago
## Wider Face Val Model|Easy Set|Medium Set|Hard Set ------|--------|----------|-------- libfacedetection v1(caffe)|0.65 |0.5 |0.233 libfacedetection v2(caffe)|0.714 |0.585 |0.306 Retinaface-Mobilenet-0.25 (Mxnet) |0.745|0.553|0.232 version-slim-320|0.77 |0.671 |0.395 version-RFB-320|0.787 |0.698 |0.438 yoloface-500k-320|0.728|0.682|0.431|
rocauc · 6 years ago
Thanks, I see the table. Are the source datasets available for creating additional benchmarks?
rjeli · 6 years ago
Wow, 100MFlop’s. That could run real time on a $5 dsp.
chvid · 6 years ago
I think it is very cool.

Trying to think of some applications for this. For example one could create a mechanism that watched people entering and exiting a shop providing the shop owner more quantitative data that he could use to optimize his sales.

Or you could have it watch a soccer game. Generating all sorts of data on how the game went.

All on relative cheap piece of hardware.

mycall · 6 years ago
Entering/exiting buses for automatic passenger counters is more important than ever now. Being able to broadcast GTFS-Occupancy in real-time when only 50% (or less) of the bus can be filled with passengers, is a real issue transit is facing today.
andrewnc · 6 years ago
layoutIfNeeded · 6 years ago
Wew, 500kb is ultra-light nowadays. I wonder how much space would the original Viola-Jones face detector take.
srg0 · 6 years ago
Check out https://github.com/opencv/opencv/tree/master/data/haarcascad...

Plain-text XML for the frontal face detector is 912 KB. 132 KB gzipped. It should be smaller in binary.

sp332 · 6 years ago
"Bflops"? I'm guessing this is a measure of the total processing power needed, in billions of floating-point ops, and not a measure of operations per second?
rjeli · 6 years ago
Yes exactly, FLOPS vs FLOPs/FLOP’s, there’s no unawkward way to write it but it’s almost always obvious from context.
anigbrowl · 6 years ago
It looks strange because most people write gigaflops/gflops.