6

It would be trained to recognize despite distortions, as a human can.

It would be trained to recognize despite distortions, as a human can.

6 comments

[–] PMYA 1 points (+1|-0)

What about fair use?

[–] jidlaph 1 points (+1|-0)

I don't think fair use is at issue here. I'm more worried about a system that can intelligently recognize political dissent and snuff it out before it even begins.

[–] PMYA 1 points (+1|-0)

A huuuuuge percentage of youtube videos have copyrighted material in them, and a lot of them come under fair use. There is also the instance of copyrighted music or something being uploaded, but the owner wants it to stay up for exposure. Youtube is already a huge money drain though, so I doubt Google would waste resources crawling through god knows how many hours of video that gets uploaded every day.

I don't think political dissent is an issue. The level of complexity involved in automating video removals for political reasons is ridiculous. By the time anyone managed to do that, I doubt Youtube would even still be a thing. It would at the very least be in a different format than it is now. If there was enough manpower to watch every second of footage uploaded to Youtube and report it for political dissent, how many of those reports would be inaccurate? I am guessing it would be a lot of them, so I highly doubt we would be able to get a machine to do it.

[–] phoxy [OP] 1 points (+1|-0)

Those rules will have to be taught as well, and if they can't be it would prevent full automation of the flagging process.

Length based fair use should be teachable, but isn't there an Educational fair use clause which allows use of the whole work? That would be very difficult discern from other uses.

[–] PMYA 1 points (+1|-0)

Fair use doesn't have to be length based. I could put Scarface on Youtube in its entirety and fart all the way through it, and argue that it comes under fair use because I am critiquing or making a parody. It doesn't have to be good or even make sense, if I can argue that it should be classed as fair use, nobody can remove it. I'm not sure about educational fair use though, don't know anything about that.

Even if you had something capable of flagging things with a decent amount of accuracy, what would it use as a database to compare Youtube videos to and decide if something should be removed? It would need to perform a reverse video search of sorts.

[–] phoxy [OP] 1 points (+1|-0)

You make good points. Fair use might torpedo the idea for the forseeable future.

The way categorization by neural network would work is it would "watch" the video and spit out a high dimensional vector (1000 dimensions for example). Similar content would have similar vectors, like content based hashing but with each dimension of the vector having a semantic meaning. The network might learn to recognize comedy or length or actors as a dimension.

If the copyright holder submits the work as a reference, similar videos can be identified because they would have very similar vectors. The vectors of copied videos would cluster together in the vector space. An upload of Scarface dubbed with farts would be close to the original in many vector dimensions except for the dimensions which encode audio information. Adjusting the distance threshold from the reference vector allows you to fine tune which videos get flagged, much like it works now where you can avoid getting flagged by distorting the upload subtly.