Remix.run Logo
talkingtab 4 hours ago

Meta said the contracting "did not meet (meta's) standards". I am sure that is true. meta's "standard" is not to reveal the illegal, immoral, unethical things meta does. No matter what the harm.

Maybe a company with those standards should not get our business. Oops, no wait, maybe they mean the Friedman Doctrine standards? In that case they are entitled to do any and every thing to make a profit. No matter what the harm.

[edit: add last two sentences]

jaidhyani 3 hours ago | parent | next [-]

I used to work for Meta. I quit largely because of intense frustrations with the company. Meta has made a lot of mistakes, overlooked a lot of harms, and made a lot of short-sighted, selfish choices. Many things about the world are worse than they could be because of choices Meta has made.

So that when I say that they really do have a zero tolerance policy for anyone using their internal systems to violate user privacy, it's not because I'm eager to defend them. It's just true (at least, it was when I was there). There are internal systems dedicated to making sure you have access to what you need to do your job, and absolutely nothing else. All content you interact with through internal tools is monitored and logged. If you get caught trying to use whatever access your job gives you for anything other than doing your job, security immediately escorts you out of the building. This is drilled into new hires early and often. For everything Meta gets wrong, they really do take this seriously.

malfist 3 hours ago | parent | next [-]

These contractors were hired to view this data. Your defense of Meta here doesn't make sense. Meta fired them for speaking out about the data Meta collects, not because they saw the data they were hired to look at.

nradov 2 hours ago | parent [-]

Meta didn't fire individual independent contractors, they terminated a contract with a vendor. It's possible they did so because some of the vendor's employees spoke out but we don't know the real reason.

(I do think these smart glasses are super creepy and I'm not defending Meta's data collection practices.)

malfist an hour ago | parent [-]

This is some real weird defense going on here.

> but we don't know the real reason.

We know the course of events. We have brains and can reason. You really expect Meta to come out and say "Yep, we fired them because they whistleblowed"

> I'm not defending Meta's data collection practices

No but you certainly seem to be over here quibbling about epistemology in the defense of Meta

causal 3 hours ago | parent | prev | next [-]

The problem is that your comment and the one you're responding to can both be true: Just because the rules are heavily enforced does not mean the right rules are in place, starting with the fact that Meta is collecting this data to begin with.

thaumasiotes 2 hours ago | parent [-]

> starting with the fact that Meta is collecting this data to begin with.

But that can't be the problem. They're collecting the data that users send them. To avoid collecting it despite the expressed wishes of the user, they'd need to be able to recognize it as untouchable.

And recognizing the data is the exact problem that this African firm was hired to help with. What do you want Meta to do?

magicalist 2 hours ago | parent | next [-]

> To avoid collecting it despite the expressed wishes of the user, they'd need to be able to recognize it as untouchable.

> And recognizing the data is the exact problem that this African firm was hired to help with. What do you want Meta to do?

This is written as if logically exhaustive, but it misses the very obvious alternative that none of these videos should have been reviewed by a human at all (aka no reason to "recognize it as untouchable"; they're all untouchable).

If you want to get stricter and talk about collecting at all, Meta already has that solution too, by leaving the video in the user's camera roll. Let the user manually add the video to the Meta AI app or whatever if they want to share it with others there.

thaumasiotes 2 hours ago | parent [-]

> This is written as if logically exhaustive, but it misses the very obvious alternative that none of these videos should have been reviewed by a human at all (aka no reason to "recognize it as untouchable"; they're all untouchable).

No, taking that approach would mean that when someone sends you data that you aren't supposed to collect, you collect it anyway. This is the opposite of what was suggested above.

magicalist 2 hours ago | parent [-]

> No, taking that approach would mean that when someone sends you data that you aren't supposed to collect, you collect it anyway. This is the opposite of what was suggested above.

That was in reference to the original story, that human annotation is happening on videos that no one knew were getting reviewed. If you want to talk about not collecting at all, well:

> If you want to get stricter and talk about collecting at all, Meta already has that solution too, by leaving the video in the user's camera roll. Let the user manually add the video to the Meta AI app or whatever if they want to share it with others there.

DrewADesign 2 hours ago | parent | prev [-]

Ok, let’s see that consent form and how explicitly it states that random call center people will possibly look at anything you record. I’ll bet you a crisp $50 it was a form designed to be as click-through-worthy as possible, being sure to not trigger the “wait, should I do this?” reflex in users, and also not loudly disclosing that you could still use the device without agreeing, if you even can, while still technically “””disclosing””” this information. The tech world has turned consent into a fucking joke.

thaumasiotes 13 minutes ago | parent [-]

I can't say anything about the consent form. The privacy policy for the glasses is here: https://www.meta.com/legal/ray-ban-stories/facebook-view-pri...

It incorporates by reference the general Facebook privacy policy. The relevant subsection is here: https://www.facebook.com/privacy/policy?subpage=4.subpage.12...

Facebook reserves the right to share any information they have about you with their contractors, for purposes including but not limited to:

- investigating suspicious activity

- improving the functionality of their products

- providing technical infrastructure services

- analyzing how their products are used

- conducting research

advisedwang 2 hours ago | parent | prev | next [-]

There's no allegation that these workers abused their access. The allegation is that their routine work reviewing footage included private content. The revelation is that USERS are using meta glasses non-consensually.

rkagerer 3 hours ago | parent | prev | next [-]

Many things about the world are worse than they could be because of choices Meta has made.

If Facebook were designed with a different set of incentives that prioritized the user, fostered positive engagement, and better respected individual's privacy and data sovereignty - setting a better standard for the whole industry - I feel there wouldn't be all this fuss today about banning social media accounts.

bilekas 3 hours ago | parent [-]

It's likely they wouldn't be as profitable too though, and their mandate to shareholders is to make number go up.

nullsanity 2 hours ago | parent [-]

[dead]

red_admiral 2 hours ago | parent | prev | next [-]

Indeed, on this one point, Meta has higher standards than the NSA used to - Snowden mentioned that employees tracked their current wives/girlfriends so often it unofficially got the codename LOVEINT.

Same for "Meta reads your E2E whatsapp messages". Meta does many things, is probably massively net negative for civilisation, but it doesn't do that.

magicalist an hour ago | parent [-]

It's kind of weird to have a subthread about "Meta doesn't do these other unrelated things" in a thread about a thing Meta is doing.

They don't boil live kittens either, I believe. Doesn't seem relevant though.

cozzyd 2 hours ago | parent | prev | next [-]

Anecdotal of course, but I heard that this wasn't at all the case circa 2006 and that (then) FB employees would routinely read private messages and such. Obviously it wasn't a big company yet and probably didn't have those policies yet... (clearly the policies are there for a reason...)

bombcar 2 hours ago | parent [-]

That’s my recollection too - there were some high profile cases and so institutional safeguards were established. They very well may be at the forefront of it - however, it’s a side issue to what’s being discussed.

thunderfork 2 hours ago | parent | prev | next [-]

As someone who worked for a contractor which had Meta as a client, I disagree.

All advertiser support agents were given super-read on all profiles & pages, and I never once observed a CSR being questioned on their use of this access in any way.

bombcar 2 hours ago | parent [-]

It’s often the case that employees are much more locked down than contractors, simply because the company is more liable for employee actions.

keybored 2 hours ago | parent | prev | next [-]

> I used to work for Meta. I quit largely because of intense frustrations with the company. Meta has made a lot of mistakes, overlooked a lot of harms, and made a lot of short-sighted, selfish choices. Many things about the world are worse than they could be because of choices Meta has made.

When did FaceBook make the world not-worse?

iJohnDoe 2 hours ago | parent | prev | next [-]

@jaidhyani I hate to burst your bubble, but there are major privacy violations here.

https://news.ycombinator.com/item?id=47226756

iJohnDoe 2 hours ago | parent | prev | next [-]

@jaidhyani I hate to burst your bubble, but there are major privacy violations here.

https://news.ycombinator.com/item?id=47225130

magicalist 3 hours ago | parent | prev | next [-]

[dead]

2ndorderthought 3 hours ago | parent | prev | next [-]

Yea but no. Meta is a defense contractor that hires out to 3rd parties exactly to do this. so you guys don't get to do that, but a lot of other people are. I hope that helped you sleep at night while you were there. But yea, it all gets bought and sold at the end of the day.

The irony is meta wants to implement verification to protect kids. Meanwhile it's doing everything it can to exploit them most at every single level for profit and for the love of the game. Billions of dollars, the world's most advanced computers all dedicated for it

3 hours ago | parent [-]
[deleted]
bathtub365 3 hours ago | parent [-]

Meta and their employees have spent years breaking the public’s trust over and over again. Why should we trust anything they say?

deaux 2 hours ago | parent | prev [-]

You're still on the koolaid, as many replies here accurately point out. Saying it's not because you're eager to defend them is lying to yourself, because you're smart enough to think of most of these replies yourself. Primarily the fact that these are contractors whose entire job is to watch smart glasses footage and the point your bringing up - even if we believe it at face value - is completely irrelevant to this post.

If you truly want to atone for your sins, you have a long way to go. I don't blame you for having worked there, I've worked at places that are only a little better than Meta (which is hard considering Meta is at the absolute bottom of the entire ladder, including Peter Thiel companies, thanks to Meta's sheer scale of carnage). But its time to completely come to terms with the reality, rather than stopping halfway to try and feel better about your resume.

deepsquirrelnet 3 hours ago | parent | prev | next [-]

> At the time of the publication, Meta admitted subcontracted workers might sometimes review content filmed on its smart glasses when people shared it with Meta AI.

They just got fired for "piercing the veil". They committed the sin of bringing attention to the invasion of privacy.

alistairSH 3 hours ago | parent [-]

Were/are video recordings from the glasses advertised as being E2E encrypted?

Mostly, I'm just surprised that anybody would be naive enough to take a camera provided by Facebook into a sexual encounter and expect anything else.

ninth_ant 2 hours ago | parent [-]

If you don’t disable the glasses they could continue to share content. The article describes the glasses being left on a dresser and then sharing content of people without their consent, which could easily parallel into showing a sexual encounter or other privacy-sensitive scenarios.

alistairSH 2 hours ago | parent [-]

Sure, and the same is true with my iPhone or my Olympus. Except the former encrypts the video and the latter isn't internet-connected.

The problem here (other than Meta being Meta) is people assuming Meta isn't permanently operating in bad faith. I'm just surprised anybody into tech to the extent they'd buy first-gen VR glasses would be surprised at Meta doing Meta things. That's all, I guess.

burnte 3 hours ago | parent | prev | next [-]

Yeah, why the hell is Meta wa5tching people's videos either? Why PAY a company to invade our privacy and watch our videos? It's flipping BIZARRE.

stingraycharles 3 hours ago | parent [-]

Isn’t that obvious from the article? They’re labeling content for training AIs, something which is happening all over the world constantly.

2ndorderthought 3 hours ago | parent | next [-]

Yep gotta bake in that personal data into generative models so it can be reproduced later for profit.

woodson 3 hours ago | parent [-]

Why generative? Or has it been decided that only generative models are “AI”?

throwpoaster 3 hours ago | parent [-]

What kind of model "reproduce"s things later for profit that is not generative?

Ritewut 3 hours ago | parent [-]

Surveillance models.

jmye 2 hours ago | parent | prev [-]

And then people are shocked that no one wants the data centers for this shit built in their backyard.

stingraycharles 3 hours ago | parent | prev | next [-]

Unfortunately in today’s world where organizations are larger than many a country’s GDP, they really only have to face responsibility towards shareholders and maximizing profits is the thing they usually care about.

throwpoaster 3 hours ago | parent | prev | next [-]

That's not what the Friedman Doctrine is, technically. It is that management should obey moral, ethical, and legal frameworks in the operation of the business for the benefit of its investors; and specifically NOT take actions which are outside of that narrow scope.

Avicebron 2 hours ago | parent [-]

Does that include trying to influence moral, ethical, and legal frameworks to the benefit of the investors as well? Because if it does it is kind of a moot point.

throwpoaster an hour ago | parent [-]

Yes, although as the Koch Brothers point out in their book: you have to play by the rules that exist, not the rules you want.

If you read, eg, Buffet, he makes the point that a manager donating to a political cause, whether the Heritage Foundation or, God forbid, something as far right as the SPLC, makes that donation with money that otherwise accrues to the shareholders. The manager therefore creates an agency problem, where he might pursue his own interests at the expense of the owners.

If they are aligned, the manager can retain the earnings and create a dividend for the owners, such that they can then make the donation directly. If they are not aligned with the owners, they are redistributing wealth.

I am not surprised that the Left advocates for backdoor wealth redistribution, but I would prefer they be honest about it.

Avicebron 38 minutes ago | parent [-]

> I am not surprised that the Left advocates for backdoor wealth redistribution, but I would prefer they be honest about it.

I'm pretty sure it's not just the Left team that advocates for bribes (sorry lobbying) to politicians. I don't think that's a very commonly held understanding of wealth redistribution either...but this argument you present isn't very coherent which is somewhat expected so I guess keep on keeping on..

prepend 3 hours ago | parent | prev [-]

Is it illegal or immoral? Having Meta review this material has to be approved by users and has their consent.

There was an example in the article where a user’s glasses kept recording the user’s wife after he took them off. That’s bad but on the user, not Facebook.

Seems similar to a situation where someone takes nudes of someone without their consent and then sends them off to a lab to be printed. The lab isn’t doing anything illegal or unethical printing them when they ask the user “are these legal” and the user replies “yes.” Unless you want to stop photo printers from ever printing nudes, I think the responsibility is on the user, not the firm.

msh 3 hours ago | parent [-]

Is there explicit approval? Or is it buried in the legal agreements?

throwpoaster 3 hours ago | parent [-]

Legal agreements are explicit.

msh 2 hours ago | parent [-]

Lol