> LinkedIn users attempting identity verification may be unknowingly handing sensitive personal data to Persona Identities Inc., a company that distributes information to government agencies, credit bureaus, utilities, and mobile providers.
^ Link from a LinkedIn page I found on a Kagi search.
I can view some LinkedIn pages but not others without logging in.
Even though I’ve never posted to LinkedIn it only use it as a public résumé, my account was flagged as needing identity verification. I’m pretty sure this happened a year or two ago when I changed my email address from one domain I owned to another domain I owned.
I’ve never been able to log in since then, and there is no support path. The only available way past it is to simply submit all the info to Persona.
(Edit: I meant to say PR, not issue...)
I've seen that before on other issue comment threads. The repo owner says "Hey everyone, if you want an issue fixed, please upvote the issue with a thumbs up". And many people don't read that, and instead post "Please fix this" comments without giving a thumbs-up to the issue. So, 1) the repo owner doesn't get to use the "sort issues by # of thumbs-up reactions" to see the priority of that issue, and 2) everyone who has subscribed to the issue gets spammed with a message that's useless to them.
Since nearly all the new comments had become "me too"-style comments, which should have just been a thumbs-up on a previous comment in order to reduce spam, I feel like locking the issue thread was the right move at that point, to stop people from receiving yet more unnecessary email in their already-overflowing inboxes.
I am reading all pings from GitHub on VS Code and this was just turning into a stream of spam that wasn't adding much new information.
The rationale I suppose is those customers what to be more careful with code that was contributed by AI.
HTH
Maybe those customers should just be more selective with the people they allow to contribute to their project?
Also, this kind of message doesn't even bring valuable info: it doesn't explain how the AI was used (could be 99% vibe-coding, or just a quick "Please review current changes" + minor fixes at the end?), which model was used, etc. Like other commenters here I can't see this as anything else than a marketing push for Copilot.
Don't take it personally though, you are probably not the one that should be taking the heat since the change was directly pushed by your product manager.