slowmovintarget a day ago

Which implies that elsewhere...

There are some consumer protections that I really do wish we imported into the U.S., especially food safety and chemical usage. Too much regulatory capture for that, though.

  • benreesman a day ago

    The FTC under Lina Khan’s enlightened, benevolent leadership seems anything but captured (though I agree with you that this is atypical): she’s the #1 enemy of everyone with a family office for a reason and if they don’t manage to throw her out, she seems to just be warming up. Everything from search monopoly to prescription drug price negotiation is on its back foot for the first time in decades.

  • WhyNotHugo 15 hours ago

    > There are some consumer protections that I really do wish we imported into the U.S., especially food safety and chemical usage.

    It's always shocking when you find that some brand-name product tastes different in the EU because an ingredient used elsewhere is not considered apt for human consumption here.

    After living a few years in The Netherlands, a lot of Argentinian candy has a faint taste of something kinda chemical. A few others who've moved here expressed perceiving the same flavour.

    My guess so far is that living in Argentina we're continuously exposed to whatever this ingredient and become desensitised to it. When we live abroad for enough years, we become sensitive to it again.

  • rsynnott 21 hours ago

    I think people are too quick to blame regulatory capture, tbh. Today, the US only really has one world-class regulator, which sets the standard: the SEC (the FTC is more active than is used to be, but still some ways from its glory days). With the result that one of the US's two political parties is now gunning for it, and depending on the outcome of the next election it may end up neutered.

    The EU is structurally difficult to effectively, ah, lobby. This, in my view, is the big differentiator; if you're a company who wants to avoid inconvenient regulations, you don't just have to compromise one political party. Unless things are very, very close, you'd have to get to, at least, tens of national parties, plus a decent bloc of MEPs, to really screw things up.

  • zaptrem a day ago

    Why do you care if someone trains an AI on content you have chosen to post publicly (LinkedIn profile/posts)? I’d understand if it was your DMs or something but this stuff is no secret.

    • ManBeardPc a day ago

      Posting images, articles and other content doesn’t grant everyone the right to use it for every purpose. Especially not to republishing it partially under the excuse a machine is doing it. It’s just not the same as someone getting inspired by it or citing it.

      Automatically doing something is a whole other quality from a person doing it. Police watching a protest is fine, police filming it or documenting all participants via face recognition is forbidden (at least here).

      • zaptrem a day ago

        I understand and completely agree republishing content (even altered) isn't cool, and I also agree government use of technology for mass surveillance is incompatible with our idea of democratic/open societies. However, in the case of LinkedIn posts you have already given the ideas behind your content to the world (and in this case specifically and explicitly LinkedIn) for free.

        I've said this before, but it's sad how quickly this community swapped from being champions of the free exchange and use of information for the betterment of humanity to gleefully stomping on an incredible and beautiful new technology because someone else might make money off it. Reminds me of the Judgement of Solomon [1] (people would rather kill the whole technology and all the incredible things that may come with it then miss out on "my cut!", "my cut!" even if it's a single LinkedIn post in a corpus of billions)

        https://en.wikipedia.org/wiki/Judgement_of_Solomon

        • ants_everywhere a day ago

          > government use of technology for mass surveillance

          Well, for context, when the US wanted to create a mass surveillance apparatus in the 1990s, they funded the tech sector. Now the tech sector both does the surveillance and fuels the economy. The surveillance info is fed back to the government in various ways and is increasingly used for things like military targeting.

          Generally speaking, any information you give to a private company you should assume will be used by your government. Either directly via subpoena/tapped data lines/etc, or indirectly via AI services provided by these companies to the government.

        • batch12 a day ago

          The Judgment of Solomon is a bad analogy as the mother is willing to give her baby away to save it. You've reversed the roles and are comparing those who want control of their data to the woman who never had a claim to the child to begin with.

          • zaptrem a day ago

            I see the open weights/source model community as the ones willing to give the baby away.

        • _DeadFred_ a day ago

          Has this community done a complete 180 randomly out of the blue, or is it a reflection of how the new technology is being used? And if this previously ultra friendly community did a 180, imagine the feelings in the general public that never had the friendly attitude.

          • CaptainFever a day ago

            > A reflection of how the new technology is being used

            This statement can be interpreted as:

            1. Some AI is bad (e.g. "some AI take jobs"), therefore LinkedIn's AI is bad. This is an invalid argument, so we can ignore this.

            2. All AI is bad, therefore LinkedIn's AI is bad. Valid, but probably unsound, because intuitively this would mean saying "AI used to fight cancer is bad". So we can ignore this too.

            3. LinkedIn's AI is bad by itself. Why? What is LinkedIn using it for? Is it just a chatbot? There's no risk of obsolete careers from that. Is it a privacy issue? It is public data. Is it because it makes money? What's wrong with that? Is it simply a fear of the new? I think so, but that's just my uncharitable opinion.

            Please elaborate.

        • allturtles 16 hours ago

          > incredible and beautiful new technology

          To me, one could call this technology incredible and beautiful in the same way that a detached observer might use those terms to describe a hurricane bearing down on my hometown. In the Kantian sense it is more "sublime"[0] rather than "beautiful", and yet the sublime is experienced as such only if it can be viewed from a position of safety; but there appears to be no safe place from AI.

          [0]: https://1000wordphilosophy.com/2023/11/11/kants-theory-of-th...

        • potamic a day ago

          > how quickly this community swapped from being champions of the free exchange and use of information for the betterment of humanity to gleefully stomping on an incredible and beautiful new technology

          You're assuming this community is a monoculture. I'm sure you'll see people on various sides, from ardent free exchange advocates to die hard copyright supporters.

        • soco a day ago

          Maybe I don't understand your comment so please help me: government use of surveillance is bad, and private use of surveillance is acceptable? So it's bad if an institution over which I could have some influence uses surveillance, but if it's a random company with hidden agendas that's totally fine?

        • Barrin92 a day ago

          >I also agree government use of technology for mass surveillance is incompatible with our idea of democratic/open societies.

          well, corporate use of technology for mass surveillance is equally incompatible with the idea of a democratic society, which is why the EU imposes limits on what LInkedIn can do with your data, and thank God for it.

          Free exchange of information is being able to access a textbook at the library, not a corporate behemoth vacuuming people's personal information so they can sell them to the highest bidder to put you under surveillance the next time you attempt to switch a job or send you more ads. Betterment of humanity? Blink twice if a LinkedIn PR person is holding you at gunpoint

          • zaptrem a day ago

            > Betterment of humanity? Blink twice if a LinkedIn PR person is holding you at gunpoint

            I was referring to Generative AI in general, this use-case is quite boring.

            > a corporate behemoth vacuuming people's personal information so they can sell them to the highest bidder

            Do people not use LinkedIn to explicitly signal to the world that they're looking for a job? Why does it matter how that information is being delivered? If you don't want the world to know you're looking for a job simply don't update LinkedIn?

            > corporate use of technology for mass surveillance is equally incompatible with the idea of a democratic society

            The argument against government power/surveillance is that they have a monopoly on it and may use their power to hurt people. It is good to legally protect sensitive information like health data from advertisers, but in this case you can, again, simply not use LinkedIn. What difference does it make if the info is collected by a company looking for new hires, a third party analytics company working on behalf of them, or LinkedIn itself working on behalf of them? It's not private data.

            • sweeter a day ago

              "just dont use LinkedIn" is such a narrow minded thing to say. How do you feasibly expect people to exist in society without interacting with any of these systems and corporations? if its not LinkedIn its Indeed or w/e else. They all collect data and most of them are pumping it into some kind of LLM or behavioral analysis algo. That is not functionally different than the argument for the Government doing it, except for that the Gov has a monopoly on violence.

              This applies for pretty much everything in our daily life like banking, shopping etc... "just don't interact" is such a useless nothing-burger that side-steps the problem entirely. You can "solve" all societal problems by becoming a hermit, moving to the woods and living off the land... but that is not a functional or reasonable thing to do for 99.99% of people. Its baffling especially when the reasonable solution is simply having a bare minimum standards of protection across the board, which many countries already implement to great effect.

              • tourmalinetaco a day ago

                I’ve never gotten a job via LinkedIn, and neither has my wife. We both got our entire job history due to connections already formed through college. In fact I can’t name a single genuine offer that came from LinkedIn/Indeed, let alone that depended on me having an account. People have been getting great paying jobs for almost the entirety of industrialized society without LinkedIn. Saying “don’t interact” with LinkedIn, especially if you disagree with their “you are the product” mentality, is a fairly realistic stance.

            • ants_everywhere a day ago

              > ... surveillance is that they have a monopoly on it

              Wait what, this isn't even close to being true

              > government power ... is that they have a monopoly on it

              This is only true by definition. McDonalds also has a monopoly on serving McDonalds food.

              > use their power to hurt people

              This is universally true of anyone with power. The difference between the government and other powerful organizations is that the government has a universal feedback mechanism via voting. In a democracy everyone gets a say in the laws that affect them.

              • zaptrem an hour ago

                I don't have to use LinkedIn and they can't throw me in jail or worse. Most people have no choice but to use their government (yeah I guess I could move but most cannot and it assumes there's another country that would take you).

          • CaptainFever a day ago

            > Free exchange of information is being able to access a textbook at the library, not a corporate behemoth vacuuming people's personal information so they can sell them to the highest bidder to put you under surveillance the next time you attempt to switch a job or send you more ads. Betterment of humanity? Blink twice if a LinkedIn PR person is holding you at gunpoint

            This sort of language is unproductive and causes further division. It's just buzzwords hoping to evoke a certain emotion in the reader. I'm being genuine when I say the following:

            > corporate behemoth

            Why is this bad? (Steelman: it isn't bad by itself, but it makes the other things worse.)

            > vacuuming

            Evokes a certain image, but it is not true. It is copied.

            > personal information

            True, but note that it is public in this case. Why is this bad? What's so special about personal information (I'm interpreting this as PII specifically)? (Steelman: it can be used to track you; see "under surveillance the next time you attempt to switch a job")

            > sell them to the highest bidder

            Why is this bad? (Steelman: it feels unfair, to have contributed something without compensation; Counterpoint: access to LinkedIn itself is your compensation, such fairness is a subjective feeling, compensation is for example not needed if you reuse GPL code in a commercial context)

            > under surveillance the next time you attempt to switch a job

            Is this true? If this is true, I can see why LinkedIn AI would be bad, but only for this specifically. If this becomes false, then it's no longer bad. I doubt it is true, however, since this is not generative AI, which they're likely focusing on.

            > send you more ads

            This has nothing to do with generative AI, which I assume we're talking about.

            > Betterment of humanity? Blink twice if a LinkedIn PR person is holding you at gunpoint

            Not a useful statement.

          • csallen a day ago

            > not a corporate behemoth vacuuming people's personal information so they can sell them to the highest bidder to put you under surveillance the next time you attempt to switch a job or send you more ads

            I genuinely don't see what the problem is.

            I rarely post updates on LinkedIn. When I do, they're updates that are intended to be broadcast to the public. If some execs at LinkedIn are smart enough to find a way to profit off the back of that, why should I be upset about that? Why are you upset about it?

            • tourmalinetaco a day ago

              Because they somehow thought a free service was there to benefit them, and not generate revenue.

      • AStonesThrow a day ago

        Well, look at it this way: you gave stuff to LinkedIn.

        Whatever their terms say, they're storing your stuff, they are serving your stuff, and they've reserved the right to extract value from your stuff in perpetuity, according to your agreements when you signed up, when you posted, and when they updated. Doesn't matter about your privacy settings, because it all happened on LI.

        I mean, you can delete the stuff you posted if you don't want future AIs trained? Delete your whole account if you didn't like LinkedIn messing with it?

        (I could likewise say this about MS Windows, Apples or anything: if you don't want someone to have your stuff, give it to someone else, or don't give at all?)

        But in the end, you voluntarily gave it to them, because it was free, but you are the product, and not the artist.

        • gomerspiles a day ago

          Microsoft probably owns the physical media from before LinkedIn was acquired so by your physical ownership logic they can keep using all the data you have "deleted" and ignore your new opt outs on all those backups..

          The point of making legislation is to have things to enforce at times like buy outs to say things there is no reasonable way to enforce our expectation that our 2FA numbers are not abused by this new buyer so the buyout can not continue.

          Maybe you don't need a job social network, but presuming you do you have no way of knowing what org will own the physical media of the one you pick today unless it is in a country with competence, I.e. not the US.

        • ManBeardPc a day ago

          I partially agree, you agree to the terms of service. However one should also not forget that LinkedIn is quite dominant and a very big player. For many businesses it is simply not an option to not be on their platform, unless they can afford to lose potential customers and employees.

          Rules change for monopolies and oligopolies, and for a good reason IMHO. LinkedIn belongs to Microsoft, so does GitHub. Don’t forget Windows, Azure, Office, Visual Studio and a long list of other products. They want to take your data from all possible sources and if you just point to the TOS alone this would be totally valid. But we have to look at the bigger picture and already do so in other areas, for example GDPR.

    • rsynnott 21 hours ago

      If you're okay with it, affirmatively consent to it. For the rest of us, out-by-default seems like a reasonable basis.

      • zaptrem 10 hours ago

        By default you do not have a LinkedIn account.

    • tempodox a day ago

      > this stuff is no secret.

      “No secret” from a human reading it is something categorically different from a machine ingesting it en masse and retaining it forever.

      • zaptrem a day ago

        Should we kill the Internet Archive too?

    • mystified5016 a day ago

      Yeah, everyone should be allowed to cut down trees on public property for firewood, or dump their trash in public parks!

      If they didn't want these resources to be exploited, they shouldn't make them publicly accessible!

      • zaptrem a day ago

        Cutting down a tree implies the tree is no longer there and cannot be used by others. In this case, the content is still there, unaltered.

      • gdhkgdhkvff a day ago

        Are we back to the “you wouldn’t download a car” pointof things?

      • hn_throwaway_99 a day ago

        This is probably the worst analogy I've read this year.

rchaud 17 hours ago

> .. End of July when X automatically enabled the training of its Grok AI on all its users – European accounts included. Just a few days after the launch, consumer organizations filed a formal privacy complaint with the Irish Data Protection Commission (DPC)[...]. The Irish Court has now dropped the privacy case against X as the platform agreed to permanently halt the collection of EU users' personal data [...].

Why drop the case? A large amount of data was still collected, was it not? I'm sure X is happy to agree to not collecting data in the future, seeing as how they've already collected, what, 18 years of data already?

Kim_Bruning a day ago

Of course this then later leads to: "Linkedin AI has non-European bias"

I'm of two minds.

  • makeitdouble a day ago

    This would be a bigger issue if Linkedin cared a lot about EU users in the first place.

    If tomorrow a user facing feature needed to be radically different between US and EU users, I think they'd also delay EU rollout until it reaches an actionable priority, so it's kinda par for the course.

  • smdyc1 a day ago

    Exactly what i was thinking.

JSDevOps a day ago

How fucked In the head do you have to be to train ANY AI on LinkedIn Data.

  • yarg a day ago

    It could be used for all sorts of things:

        You could use it to indicate the fit between employees and companies;
        You could use it to detect lies and exaggerations in CVs;
        You could use it to estimate when employees are likely to be considering seeking new opportunities.
    
    There's significant potential for business relevant applications.
    • bastawhiz a day ago

      I'm looking forward to getting fired because my employer thinks I'm considering new opportunities because they used LinkedIn AI

      • yarg a day ago

        That's cool, but when decent companies think that legitimately useful people are considering resigning they often don't fire but instead offer a raise.

      • chrsw a day ago

        Don't worry, people are hard at work making sure these AI systems are "aligned".

    • wkat4242 a day ago

      But linkedin is far from accurate information. It's full of hollow mindless corporate PR and trumped up CVs.

  • bqmjjx0kac a day ago

    I think it depends on what your objective is. Like if you want to simulate the LinkedIn experience, it's natural to train on LinkedIn data.

  • specproc 16 hours ago

    Exactly. Who wakes up in the morning and thinks, "You know what the world needs? An LLM that writes linkedin influencer-style posts".

    • mewpmewp2 15 hours ago

      The linkedin influencers probably?

  • Havoc a day ago

    HR and marketing will love it.

    Make of that what you will

  • Brajeshwar a day ago

    Unfortunately, the bitter news is that LinkedIn works for 98.42% of people.

  • playingalong a day ago

    Well, not a general purpose one, but...

    knowing LI doesn't have any NSFW content, is full of marketing content (both corporate marketing and self boasting individuals), tends to prefer positive signals (all projects succeed, etc.), is mostly done in English, and so on...

    ... There is clearly a market for text generation in this language bubble. Think of all the internal and external communication in your $BigCorp. That has immediate use not only in marketing, but also HR, recruitment, company policing, etc. Your next company town hall can be prepared and led (!) by this thing.

    • bastawhiz a day ago

      Half of LinkedIn posts are already written by chatgpt, you don't need a model trained on the output of another model

      • jart a day ago

        You just described most open source models.

  • artursapek a day ago

    maybe someone needs an AI that generates smug self-congratulatory word salad

    • yarg a day ago

      No, but the ability to probabilistically detect it could be used to build useful filtering and prioritisation functions.

    • rsynnott 21 hours ago

      ... I mean, isn't that all of them?

  • gedy a day ago

    Training is one thing, but that feature to have AI help you with your LI posts seems psychotic. Generic generated slop for what exactly.

wodenokoto a day ago

What is a European user anyway.

Someone who created their account from EU IP-address, someone who says they live in EU, someone who says the work from a company (division) based in EU?

left-struck a day ago

I wonder how this works for Europeans abroad. More to the point, as an Australian can I trick these tech companies into thinking I’m a European because my government won’t protect me.

Chris_Newton a day ago

In the UK, I found the relevant setting was present (and had been turned on) for my profile when this came to my attention last week. Curiously, I can no longer find that setting now.

Given the UK’s privacy and data protection environment is still largely the same as the EU’s, I wonder whether this was an oversight.

Brajeshwar a day ago

It might sound naive, but how does LinkedIn confirm and be sure that someone is a European user?

amarcheschi a day ago

Yup, I noticed this a few days ago in some subreddit like /r/assholedesign, I think a few months ago we had a similar feature on instagram and perhaps fb, I don't know if it's still active in EU on those meta products

bastard_op a day ago

Just imagine if we had the same privacy protections as the EU in the US.

  • antegamisou a day ago

    Well one can't have everything.

    The average webdev in many countries of Europe earns pennies and cost of living being still abysmal. Meanwhile one can only be breathing in the USA with a CS degree and easily make quite a lot, more than enough to get by.

    • rcbdev a day ago

      In the city center of Vienna rent for my three room apartment is 600 bucks. When I used to live in the U.S. I earned more for sure, but the quality of life and cost of living was such crap in comparison that it could never make it worth it.

      I don't really mind that I earn less here.

    • wkat4242 a day ago

      Development makes decent money here. Not the 250k$ you'd get in silicon valley but the costs of living are also way lower here. It's definitely one of the better classes of jobs.

ilrwbwrkhv a day ago

LinkedIn is the bottom of the barrel of the labour pool. Wonder why even train the data with any of them.

cyanydeez a day ago

Imagine an AI that's only allowed to mine the data of people to stupid to elect representatives that protect their privacy.

ein0p a day ago

I wonder how they use US data too. LinkedIn is so cringe, the value of its data in the mix is probably negative.

yazzku a day ago

Fuck LinkedIn. They should have already been sued for their exploitation of people's identities a long time ago.

zmmmmm a day ago

I know Europeans will probably be all high fiving each other and congratulating themselves on how much better their regulatory environment is.

But consider the other side of this coin : one of the biggest risks identified for AI is bias in training sets and there are actual demands that companies explicitly make their training sets as inclusive as possible to incorporate all cultures, genders, etc etc.

So if Europeans find out they are being excluded from job opportunities later on because employers are using AI tools within LinkedIn to process candidates and it simply doesn't understand the background of European candidates - will they be upset? Will they be demanding LinkedIn be fined for not including Europeans in the training set?

I am very curious how all this will play out long term as these competing tensions get worked through.

  • edelbitter a day ago

    Those "demands" for inclusive training sets are a spin attempt. And a bad one at that: It does not even address the main problem. If I refuse to be judged by a person trying to evade legal review, then that situation does not improve once they promises to make their still-hidden thought process at least as biased as the regional internet spam variety plus a few fan fiction texts for which they violated my copyrights.

    The whole idea of "you cannot sue me for my discriminatory practices, because I colluded with AI company X to hide the process" is a net negative. It does not matter which side of the coin is up, when I have to pay for it.

  • tempaccount420 a day ago

    Why can't they get that training data in another, legal way?