Security

Token

There is no security token.

Firewall

main Name
Security enabled
Stateless

Configuration

Key Value
provider security.user.provider.concrete.app_user_provider
context main
entry_point App\Security\KbinAuthenticator
user_checker App\Security\UserChecker
access_denied_handler (none)
access_denied_url (none)
authenticators
[
  "two_factor"
  "remember_me"
  "App\Security\KbinAuthenticator"
  "App\Security\FacebookAuthenticator"
  "App\Security\GoogleAuthenticator"
  "App\Security\GithubAuthenticator"
  "App\Security\KeycloakAuthenticator"
]

Listeners

Listener Duration Response
Symfony\Component\Security\Http\Firewall\ChannelListener {#723
  -map: Symfony\Component\Security\Http\AccessMap {#722 …}
  -logger: Monolog\Logger {#783 …}
  -httpPort: 80
  -httpsPort: 443
}
0.00 ms (none)
Symfony\Component\Security\Http\Firewall\ContextListener {#706
  -tokenStorage: Symfony\Component\Security\Core\Authentication\Token\Storage\TokenStorage {#1017 …}
  -sessionKey: "_security_main"
  -logger: Monolog\Logger {#783 …}
  -userProviders: Symfony\Component\DependencyInjection\Argument\RewindableGenerator {#705 …}
  -dispatcher: Symfony\Component\EventDispatcher\Debug\TraceableEventDispatcher {#747 …}
  -registered: false
  -trustResolver: Scheb\TwoFactorBundle\Security\Authentication\AuthenticationTrustResolver {#780 …}
  -sessionTrackerEnabler: Symfony\Component\Security\Core\Authentication\Token\Storage\UsageTrackingTokenStorage::enableUsageTracking(): void {#703 …}
}
0.22 ms (none)
Symfony\Component\Security\Http\Firewall\AuthenticatorManagerListener {#584
  -authenticatorManager: Symfony\Component\Security\Http\Authentication\AuthenticatorManager {#595 …}
}
0.00 ms (none)
Scheb\TwoFactorBundle\Security\Http\Firewall\TwoFactorAccessListener {#582
  -twoFactorFirewallConfig: Scheb\TwoFactorBundle\Security\TwoFactor\TwoFactorFirewallConfig {#842 …}
  -tokenStorage: Symfony\Component\Security\Core\Authentication\Token\Storage\UsageTrackingTokenStorage {#1018 …}
  -twoFactorAccessDecider: Scheb\TwoFactorBundle\Security\Authorization\TwoFactorAccessDecider {#581 …}
}
0.05 ms (none)
Symfony\Component\Security\Http\Firewall\AccessListener {#579
  -tokenStorage: Symfony\Component\Security\Core\Authentication\Token\Storage\UsageTrackingTokenStorage {#1018 …}
  -accessDecisionManager: Symfony\Component\Security\Core\Authorization\TraceableAccessDecisionManager {#937 …}
  -map: Symfony\Component\Security\Http\AccessMap {#722 …}
}
0.00 ms (none)
Symfony\Component\Security\Http\Firewall\LogoutListener {#786
  -tokenStorage: Symfony\Component\Security\Core\Authentication\Token\Storage\UsageTrackingTokenStorage {#1018 …}
  -options: [
    "csrf_parameter" => "_csrf_token"
    "csrf_token_id" => "logout"
    "logout_path" => "app_logout"
  ]
  -httpUtils: Symfony\Component\Security\Http\HttpUtils {#841 …}
  -csrfTokenManager: Symfony\Component\Security\Csrf\CsrfTokenManager {#1015 …}
  -eventDispatcher: Symfony\Component\EventDispatcher\Debug\TraceableEventDispatcher {#747 …}
}
0.00 ms (none)

Authenticators

No authenticators have been recorded. Check previous profiles on your authentication endpoint.

Access Decision

affirmative Strategy
# Voter class
1
"Symfony\Component\Security\Core\Authorization\Voter\AuthenticatedVoter"
2
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
3
"Symfony\Component\Security\Core\Authorization\Voter\RoleHierarchyVoter"
4
"Symfony\Component\Security\Core\Authorization\Voter\ExpressionVoter"
5
"App\Security\Voter\EntryCommentVoter"
6
"App\Security\Voter\EntryVoter"
7
"App\Security\Voter\MagazineVoter"
8
"App\Security\Voter\MessageThreadVoter"
9
"App\Security\Voter\MessageVoter"
10
"App\Security\Voter\NotificationVoter"
11
"App\Security\Voter\OAuth2UserConsentVoter"
12
"App\Security\Voter\PostCommentVoter"
13
"App\Security\Voter\PostVoter"
14
"App\Security\Voter\UserVoter"

Access decision log

# Result Attributes Object
1 DENIED ROLE_USER
null
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"Symfony\Component\Security\Core\Authorization\Voter\RoleHierarchyVoter"
ACCESS DENIED
"App\Security\Voter\EntryCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
2 DENIED moderate
App\Entity\Entry {#2399
  +user: Proxies\__CG__\App\Entity\User {#1966 …}
  +magazine: App\Entity\Magazine {#262
    +icon: Proxies\__CG__\App\Entity\Image {#242 …}
    +name: "RedditMigration@kbin.social"
    +title: "RedditMigration"
    +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
    +rules: null
    +subscriptionsCount: 0
    +entryCount: 230
    +entryCommentCount: 3588
    +postCount: 46
    +postCommentCount: 100
    +isAdult: false
    +customCss: null
    +lastActive: DateTime @1689732875 {#270
      date: 2023-07-19 04:14:35.0 +02:00
    }
    +markedForDeletionAt: null
    +tags: null
    +moderators: Doctrine\ORM\PersistentCollection {#233 …}
    +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
    +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
    +entries: Doctrine\ORM\PersistentCollection {#176 …}
    +posts: Doctrine\ORM\PersistentCollection {#134 …}
    +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
    +bans: Doctrine\ORM\PersistentCollection {#113 …}
    +reports: Doctrine\ORM\PersistentCollection {#99 …}
    +badges: Doctrine\ORM\PersistentCollection {#77 …}
    +logs: Doctrine\ORM\PersistentCollection {#67 …}
    +awards: Doctrine\ORM\PersistentCollection {#2434 …}
    +categories: Doctrine\ORM\PersistentCollection {#1804 …}
    -id: 19
    +apId: "RedditMigration@kbin.social"
    +apProfileId: "https://kbin.social/m/RedditMigration"
    +apPublicUrl: "https://kbin.social/m/RedditMigration"
    +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
    +apInboxUrl: "https://kbin.social/f/inbox"
    +apDomain: "kbin.social"
    +apPreferredUsername: "RedditMigration"
    +apDiscoverable: true
    +apManuallyApprovesFollowers: null
    +privateKey: null
    +publicKey: null
    +apFetchedAt: DateTime @1688009400 {#268
      date: 2023-06-29 05:30:00.0 +02:00
    }
    +apDeletedAt: null
    +apTimeoutAt: null
    +visibility: "visible             "
    +createdAt: DateTimeImmutable @1688009399 {#276
      date: 2023-06-29 05:29:59.0 +02:00
    }
  }
  +image: null
  +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
  +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
  +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
  +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
  +body: """
    A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
    \n
    I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
    \n
    > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
    \n
    > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
    \n
    > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
    \n
    > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
    \n
    > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
    \n
    Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
    \n
    > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
    """
  +type: "link"
  +lang: "en"
  +isOc: false
  +hasEmbed: false
  +commentCount: 2
  +favouriteCount: 8
  +score: 0
  +isAdult: false
  +sticky: false
  +lastActive: DateTime @1688104448 {#1828
    date: 2023-06-30 07:54:08.0 +02:00
  }
  +ip: null
  +adaAmount: 0
  +tags: null
  +mentions: null
  +comments: Doctrine\ORM\PersistentCollection {#1905 …}
  +votes: Doctrine\ORM\PersistentCollection {#1934 …}
  +reports: Doctrine\ORM\PersistentCollection {#1929 …}
  +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
  +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
  +badges: Doctrine\ORM\PersistentCollection {#2357 …}
  +children: []
  -id: 547
  -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
  -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
  +cross: false
  +upVotes: 0
  +downVotes: 0
  +ranking: 1688088743
  +visibility: "visible             "
  +apId: "https://lemmy.world/post/804780"
  +editedAt: null
  +createdAt: DateTimeImmutable @1688057743 {#2414
    date: 2023-06-29 18:55:43.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryVoter"
ACCESS DENIED
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
3 DENIED edit
App\Entity\Entry {#2399
  +user: Proxies\__CG__\App\Entity\User {#1966 …}
  +magazine: App\Entity\Magazine {#262
    +icon: Proxies\__CG__\App\Entity\Image {#242 …}
    +name: "RedditMigration@kbin.social"
    +title: "RedditMigration"
    +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
    +rules: null
    +subscriptionsCount: 0
    +entryCount: 230
    +entryCommentCount: 3588
    +postCount: 46
    +postCommentCount: 100
    +isAdult: false
    +customCss: null
    +lastActive: DateTime @1689732875 {#270
      date: 2023-07-19 04:14:35.0 +02:00
    }
    +markedForDeletionAt: null
    +tags: null
    +moderators: Doctrine\ORM\PersistentCollection {#233 …}
    +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
    +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
    +entries: Doctrine\ORM\PersistentCollection {#176 …}
    +posts: Doctrine\ORM\PersistentCollection {#134 …}
    +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
    +bans: Doctrine\ORM\PersistentCollection {#113 …}
    +reports: Doctrine\ORM\PersistentCollection {#99 …}
    +badges: Doctrine\ORM\PersistentCollection {#77 …}
    +logs: Doctrine\ORM\PersistentCollection {#67 …}
    +awards: Doctrine\ORM\PersistentCollection {#2434 …}
    +categories: Doctrine\ORM\PersistentCollection {#1804 …}
    -id: 19
    +apId: "RedditMigration@kbin.social"
    +apProfileId: "https://kbin.social/m/RedditMigration"
    +apPublicUrl: "https://kbin.social/m/RedditMigration"
    +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
    +apInboxUrl: "https://kbin.social/f/inbox"
    +apDomain: "kbin.social"
    +apPreferredUsername: "RedditMigration"
    +apDiscoverable: true
    +apManuallyApprovesFollowers: null
    +privateKey: null
    +publicKey: null
    +apFetchedAt: DateTime @1688009400 {#268
      date: 2023-06-29 05:30:00.0 +02:00
    }
    +apDeletedAt: null
    +apTimeoutAt: null
    +visibility: "visible             "
    +createdAt: DateTimeImmutable @1688009399 {#276
      date: 2023-06-29 05:29:59.0 +02:00
    }
  }
  +image: null
  +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
  +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
  +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
  +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
  +body: """
    A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
    \n
    I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
    \n
    > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
    \n
    > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
    \n
    > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
    \n
    > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
    \n
    > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
    \n
    Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
    \n
    > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
    """
  +type: "link"
  +lang: "en"
  +isOc: false
  +hasEmbed: false
  +commentCount: 2
  +favouriteCount: 8
  +score: 0
  +isAdult: false
  +sticky: false
  +lastActive: DateTime @1688104448 {#1828
    date: 2023-06-30 07:54:08.0 +02:00
  }
  +ip: null
  +adaAmount: 0
  +tags: null
  +mentions: null
  +comments: Doctrine\ORM\PersistentCollection {#1905 …}
  +votes: Doctrine\ORM\PersistentCollection {#1934 …}
  +reports: Doctrine\ORM\PersistentCollection {#1929 …}
  +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
  +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
  +badges: Doctrine\ORM\PersistentCollection {#2357 …}
  +children: []
  -id: 547
  -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
  -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
  +cross: false
  +upVotes: 0
  +downVotes: 0
  +ranking: 1688088743
  +visibility: "visible             "
  +apId: "https://lemmy.world/post/804780"
  +editedAt: null
  +createdAt: DateTimeImmutable @1688057743 {#2414
    date: 2023-06-29 18:55:43.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryVoter"
ACCESS DENIED
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
4 DENIED moderate
App\Entity\Entry {#2399
  +user: Proxies\__CG__\App\Entity\User {#1966 …}
  +magazine: App\Entity\Magazine {#262
    +icon: Proxies\__CG__\App\Entity\Image {#242 …}
    +name: "RedditMigration@kbin.social"
    +title: "RedditMigration"
    +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
    +rules: null
    +subscriptionsCount: 0
    +entryCount: 230
    +entryCommentCount: 3588
    +postCount: 46
    +postCommentCount: 100
    +isAdult: false
    +customCss: null
    +lastActive: DateTime @1689732875 {#270
      date: 2023-07-19 04:14:35.0 +02:00
    }
    +markedForDeletionAt: null
    +tags: null
    +moderators: Doctrine\ORM\PersistentCollection {#233 …}
    +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
    +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
    +entries: Doctrine\ORM\PersistentCollection {#176 …}
    +posts: Doctrine\ORM\PersistentCollection {#134 …}
    +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
    +bans: Doctrine\ORM\PersistentCollection {#113 …}
    +reports: Doctrine\ORM\PersistentCollection {#99 …}
    +badges: Doctrine\ORM\PersistentCollection {#77 …}
    +logs: Doctrine\ORM\PersistentCollection {#67 …}
    +awards: Doctrine\ORM\PersistentCollection {#2434 …}
    +categories: Doctrine\ORM\PersistentCollection {#1804 …}
    -id: 19
    +apId: "RedditMigration@kbin.social"
    +apProfileId: "https://kbin.social/m/RedditMigration"
    +apPublicUrl: "https://kbin.social/m/RedditMigration"
    +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
    +apInboxUrl: "https://kbin.social/f/inbox"
    +apDomain: "kbin.social"
    +apPreferredUsername: "RedditMigration"
    +apDiscoverable: true
    +apManuallyApprovesFollowers: null
    +privateKey: null
    +publicKey: null
    +apFetchedAt: DateTime @1688009400 {#268
      date: 2023-06-29 05:30:00.0 +02:00
    }
    +apDeletedAt: null
    +apTimeoutAt: null
    +visibility: "visible             "
    +createdAt: DateTimeImmutable @1688009399 {#276
      date: 2023-06-29 05:29:59.0 +02:00
    }
  }
  +image: null
  +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
  +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
  +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
  +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
  +body: """
    A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
    \n
    I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
    \n
    > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
    \n
    > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
    \n
    > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
    \n
    > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
    \n
    > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
    \n
    Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
    \n
    > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
    """
  +type: "link"
  +lang: "en"
  +isOc: false
  +hasEmbed: false
  +commentCount: 2
  +favouriteCount: 8
  +score: 0
  +isAdult: false
  +sticky: false
  +lastActive: DateTime @1688104448 {#1828
    date: 2023-06-30 07:54:08.0 +02:00
  }
  +ip: null
  +adaAmount: 0
  +tags: null
  +mentions: null
  +comments: Doctrine\ORM\PersistentCollection {#1905 …}
  +votes: Doctrine\ORM\PersistentCollection {#1934 …}
  +reports: Doctrine\ORM\PersistentCollection {#1929 …}
  +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
  +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
  +badges: Doctrine\ORM\PersistentCollection {#2357 …}
  +children: []
  -id: 547
  -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
  -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
  +cross: false
  +upVotes: 0
  +downVotes: 0
  +ranking: 1688088743
  +visibility: "visible             "
  +apId: "https://lemmy.world/post/804780"
  +editedAt: null
  +createdAt: DateTimeImmutable @1688057743 {#2414
    date: 2023-06-29 18:55:43.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryVoter"
ACCESS DENIED
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
5 DENIED ROLE_USER
null
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"Symfony\Component\Security\Core\Authorization\Voter\RoleHierarchyVoter"
ACCESS DENIED
"App\Security\Voter\EntryCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
6 DENIED moderate
App\Entity\EntryComment {#4077
  +user: App\Entity\User {#4025 …}
  +entry: App\Entity\Entry {#2399
    +user: Proxies\__CG__\App\Entity\User {#1966 …}
    +magazine: App\Entity\Magazine {#262
      +icon: Proxies\__CG__\App\Entity\Image {#242 …}
      +name: "RedditMigration@kbin.social"
      +title: "RedditMigration"
      +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
      +rules: null
      +subscriptionsCount: 0
      +entryCount: 230
      +entryCommentCount: 3588
      +postCount: 46
      +postCommentCount: 100
      +isAdult: false
      +customCss: null
      +lastActive: DateTime @1689732875 {#270
        date: 2023-07-19 04:14:35.0 +02:00
      }
      +markedForDeletionAt: null
      +tags: null
      +moderators: Doctrine\ORM\PersistentCollection {#233 …}
      +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
      +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
      +entries: Doctrine\ORM\PersistentCollection {#176 …}
      +posts: Doctrine\ORM\PersistentCollection {#134 …}
      +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
      +bans: Doctrine\ORM\PersistentCollection {#113 …}
      +reports: Doctrine\ORM\PersistentCollection {#99 …}
      +badges: Doctrine\ORM\PersistentCollection {#77 …}
      +logs: Doctrine\ORM\PersistentCollection {#67 …}
      +awards: Doctrine\ORM\PersistentCollection {#2434 …}
      +categories: Doctrine\ORM\PersistentCollection {#1804 …}
      -id: 19
      +apId: "RedditMigration@kbin.social"
      +apProfileId: "https://kbin.social/m/RedditMigration"
      +apPublicUrl: "https://kbin.social/m/RedditMigration"
      +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
      +apInboxUrl: "https://kbin.social/f/inbox"
      +apDomain: "kbin.social"
      +apPreferredUsername: "RedditMigration"
      +apDiscoverable: true
      +apManuallyApprovesFollowers: null
      +privateKey: null
      +publicKey: null
      +apFetchedAt: DateTime @1688009400 {#268
        date: 2023-06-29 05:30:00.0 +02:00
      }
      +apDeletedAt: null
      +apTimeoutAt: null
      +visibility: "visible             "
      +createdAt: DateTimeImmutable @1688009399 {#276
        date: 2023-06-29 05:29:59.0 +02:00
      }
    }
    +image: null
    +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
    +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
    +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
    +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
    +body: """
      A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
      \n
      I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
      \n
      > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
      \n
      > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
      \n
      > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
      \n
      > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
      \n
      > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
      \n
      Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
      \n
      > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
      """
    +type: "link"
    +lang: "en"
    +isOc: false
    +hasEmbed: false
    +commentCount: 2
    +favouriteCount: 8
    +score: 0
    +isAdult: false
    +sticky: false
    +lastActive: DateTime @1688104448 {#1828
      date: 2023-06-30 07:54:08.0 +02:00
    }
    +ip: null
    +adaAmount: 0
    +tags: null
    +mentions: null
    +comments: Doctrine\ORM\PersistentCollection {#1905 …}
    +votes: Doctrine\ORM\PersistentCollection {#1934 …}
    +reports: Doctrine\ORM\PersistentCollection {#1929 …}
    +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
    +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
    +badges: Doctrine\ORM\PersistentCollection {#2357 …}
    +children: []
    -id: 547
    -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
    -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
    +cross: false
    +upVotes: 0
    +downVotes: 0
    +ranking: 1688088743
    +visibility: "visible             "
    +apId: "https://lemmy.world/post/804780"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688057743 {#2414
      date: 2023-06-29 18:55:43.0 +02:00
    }
  }
  +magazine: App\Entity\Magazine {#262}
  +image: null
  +parent: null
  +root: null
  +body: """
    As we are on the eve of rexxit - Is there a "best" way to sabotage our posts?\n
    \n
    I suppose I see two ways of achieving this - 1) a single AI-response that we edit all posts with; or 2) actually using an AI to "reply", as in different posts which emulate the answers a human would provide but generated by AI\n
    \n
    Imo, route 2 would be more time-consuming but harder to 'prevent' from degrading the dataset from reddit's perspective?
    """
  +lang: "en"
  +isAdult: false
  +favouriteCount: 1
  +score: 0
  +lastActive: DateTime @1688104452 {#4086
    date: 2023-06-30 07:54:12.0 +02:00
  }
  +ip: null
  +tags: null
  +mentions: [
    "@NevermindNoMind@lemmy.world"
  ]
  +children: Doctrine\ORM\PersistentCollection {#4075 …}
  +nested: Doctrine\ORM\PersistentCollection {#4073 …}
  +votes: Doctrine\ORM\PersistentCollection {#4071 …}
  +reports: Doctrine\ORM\PersistentCollection {#4069 …}
  +favourites: Doctrine\ORM\PersistentCollection {#4037 …}
  +notifications: Doctrine\ORM\PersistentCollection {#4041 …}
  -id: 5540
  -bodyTs: "'1':27 '2':40,65 'achiev':25 'actual':41 'ai':31,44,62 'ai-respons':30 'answer':54 'best':12 'consum':71 'dataset':79 'degrad':77 'differ':49 'edit':35 'emul':52 'eve':6 'generat':60 'harder':73 'human':56 'imo':63 'perspect':83 'post':17,37,50 'prevent':75 'provid':58 'reddit':81 'repli':46 'respons':32 'rexxit':8 'rout':64 'sabotag':15 'see':21 'singl':29 'suppos':19 'time':70 'time-consum':69 'two':22 'use':42 'way':13,23 'would':57,66"
  +ranking: 0
  +commentCount: 0
  +upVotes: 0
  +downVotes: 0
  +visibility: "visible             "
  +apId: "https://lemmy.world/comment/632155"
  +editedAt: null
  +createdAt: DateTimeImmutable @1688058152 {#4022
    date: 2023-06-29 19:02:32.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS DENIED
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
7 DENIED edit
App\Entity\EntryComment {#4077
  +user: App\Entity\User {#4025 …}
  +entry: App\Entity\Entry {#2399
    +user: Proxies\__CG__\App\Entity\User {#1966 …}
    +magazine: App\Entity\Magazine {#262
      +icon: Proxies\__CG__\App\Entity\Image {#242 …}
      +name: "RedditMigration@kbin.social"
      +title: "RedditMigration"
      +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
      +rules: null
      +subscriptionsCount: 0
      +entryCount: 230
      +entryCommentCount: 3588
      +postCount: 46
      +postCommentCount: 100
      +isAdult: false
      +customCss: null
      +lastActive: DateTime @1689732875 {#270
        date: 2023-07-19 04:14:35.0 +02:00
      }
      +markedForDeletionAt: null
      +tags: null
      +moderators: Doctrine\ORM\PersistentCollection {#233 …}
      +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
      +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
      +entries: Doctrine\ORM\PersistentCollection {#176 …}
      +posts: Doctrine\ORM\PersistentCollection {#134 …}
      +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
      +bans: Doctrine\ORM\PersistentCollection {#113 …}
      +reports: Doctrine\ORM\PersistentCollection {#99 …}
      +badges: Doctrine\ORM\PersistentCollection {#77 …}
      +logs: Doctrine\ORM\PersistentCollection {#67 …}
      +awards: Doctrine\ORM\PersistentCollection {#2434 …}
      +categories: Doctrine\ORM\PersistentCollection {#1804 …}
      -id: 19
      +apId: "RedditMigration@kbin.social"
      +apProfileId: "https://kbin.social/m/RedditMigration"
      +apPublicUrl: "https://kbin.social/m/RedditMigration"
      +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
      +apInboxUrl: "https://kbin.social/f/inbox"
      +apDomain: "kbin.social"
      +apPreferredUsername: "RedditMigration"
      +apDiscoverable: true
      +apManuallyApprovesFollowers: null
      +privateKey: null
      +publicKey: null
      +apFetchedAt: DateTime @1688009400 {#268
        date: 2023-06-29 05:30:00.0 +02:00
      }
      +apDeletedAt: null
      +apTimeoutAt: null
      +visibility: "visible             "
      +createdAt: DateTimeImmutable @1688009399 {#276
        date: 2023-06-29 05:29:59.0 +02:00
      }
    }
    +image: null
    +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
    +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
    +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
    +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
    +body: """
      A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
      \n
      I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
      \n
      > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
      \n
      > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
      \n
      > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
      \n
      > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
      \n
      > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
      \n
      Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
      \n
      > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
      """
    +type: "link"
    +lang: "en"
    +isOc: false
    +hasEmbed: false
    +commentCount: 2
    +favouriteCount: 8
    +score: 0
    +isAdult: false
    +sticky: false
    +lastActive: DateTime @1688104448 {#1828
      date: 2023-06-30 07:54:08.0 +02:00
    }
    +ip: null
    +adaAmount: 0
    +tags: null
    +mentions: null
    +comments: Doctrine\ORM\PersistentCollection {#1905 …}
    +votes: Doctrine\ORM\PersistentCollection {#1934 …}
    +reports: Doctrine\ORM\PersistentCollection {#1929 …}
    +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
    +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
    +badges: Doctrine\ORM\PersistentCollection {#2357 …}
    +children: []
    -id: 547
    -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
    -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
    +cross: false
    +upVotes: 0
    +downVotes: 0
    +ranking: 1688088743
    +visibility: "visible             "
    +apId: "https://lemmy.world/post/804780"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688057743 {#2414
      date: 2023-06-29 18:55:43.0 +02:00
    }
  }
  +magazine: App\Entity\Magazine {#262}
  +image: null
  +parent: null
  +root: null
  +body: """
    As we are on the eve of rexxit - Is there a "best" way to sabotage our posts?\n
    \n
    I suppose I see two ways of achieving this - 1) a single AI-response that we edit all posts with; or 2) actually using an AI to "reply", as in different posts which emulate the answers a human would provide but generated by AI\n
    \n
    Imo, route 2 would be more time-consuming but harder to 'prevent' from degrading the dataset from reddit's perspective?
    """
  +lang: "en"
  +isAdult: false
  +favouriteCount: 1
  +score: 0
  +lastActive: DateTime @1688104452 {#4086
    date: 2023-06-30 07:54:12.0 +02:00
  }
  +ip: null
  +tags: null
  +mentions: [
    "@NevermindNoMind@lemmy.world"
  ]
  +children: Doctrine\ORM\PersistentCollection {#4075 …}
  +nested: Doctrine\ORM\PersistentCollection {#4073 …}
  +votes: Doctrine\ORM\PersistentCollection {#4071 …}
  +reports: Doctrine\ORM\PersistentCollection {#4069 …}
  +favourites: Doctrine\ORM\PersistentCollection {#4037 …}
  +notifications: Doctrine\ORM\PersistentCollection {#4041 …}
  -id: 5540
  -bodyTs: "'1':27 '2':40,65 'achiev':25 'actual':41 'ai':31,44,62 'ai-respons':30 'answer':54 'best':12 'consum':71 'dataset':79 'degrad':77 'differ':49 'edit':35 'emul':52 'eve':6 'generat':60 'harder':73 'human':56 'imo':63 'perspect':83 'post':17,37,50 'prevent':75 'provid':58 'reddit':81 'repli':46 'respons':32 'rexxit':8 'rout':64 'sabotag':15 'see':21 'singl':29 'suppos':19 'time':70 'time-consum':69 'two':22 'use':42 'way':13,23 'would':57,66"
  +ranking: 0
  +commentCount: 0
  +upVotes: 0
  +downVotes: 0
  +visibility: "visible             "
  +apId: "https://lemmy.world/comment/632155"
  +editedAt: null
  +createdAt: DateTimeImmutable @1688058152 {#4022
    date: 2023-06-29 19:02:32.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS DENIED
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
8 DENIED moderate
App\Entity\EntryComment {#4077
  +user: App\Entity\User {#4025 …}
  +entry: App\Entity\Entry {#2399
    +user: Proxies\__CG__\App\Entity\User {#1966 …}
    +magazine: App\Entity\Magazine {#262
      +icon: Proxies\__CG__\App\Entity\Image {#242 …}
      +name: "RedditMigration@kbin.social"
      +title: "RedditMigration"
      +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
      +rules: null
      +subscriptionsCount: 0
      +entryCount: 230
      +entryCommentCount: 3588
      +postCount: 46
      +postCommentCount: 100
      +isAdult: false
      +customCss: null
      +lastActive: DateTime @1689732875 {#270
        date: 2023-07-19 04:14:35.0 +02:00
      }
      +markedForDeletionAt: null
      +tags: null
      +moderators: Doctrine\ORM\PersistentCollection {#233 …}
      +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
      +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
      +entries: Doctrine\ORM\PersistentCollection {#176 …}
      +posts: Doctrine\ORM\PersistentCollection {#134 …}
      +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
      +bans: Doctrine\ORM\PersistentCollection {#113 …}
      +reports: Doctrine\ORM\PersistentCollection {#99 …}
      +badges: Doctrine\ORM\PersistentCollection {#77 …}
      +logs: Doctrine\ORM\PersistentCollection {#67 …}
      +awards: Doctrine\ORM\PersistentCollection {#2434 …}
      +categories: Doctrine\ORM\PersistentCollection {#1804 …}
      -id: 19
      +apId: "RedditMigration@kbin.social"
      +apProfileId: "https://kbin.social/m/RedditMigration"
      +apPublicUrl: "https://kbin.social/m/RedditMigration"
      +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
      +apInboxUrl: "https://kbin.social/f/inbox"
      +apDomain: "kbin.social"
      +apPreferredUsername: "RedditMigration"
      +apDiscoverable: true
      +apManuallyApprovesFollowers: null
      +privateKey: null
      +publicKey: null
      +apFetchedAt: DateTime @1688009400 {#268
        date: 2023-06-29 05:30:00.0 +02:00
      }
      +apDeletedAt: null
      +apTimeoutAt: null
      +visibility: "visible             "
      +createdAt: DateTimeImmutable @1688009399 {#276
        date: 2023-06-29 05:29:59.0 +02:00
      }
    }
    +image: null
    +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
    +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
    +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
    +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
    +body: """
      A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
      \n
      I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
      \n
      > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
      \n
      > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
      \n
      > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
      \n
      > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
      \n
      > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
      \n
      Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
      \n
      > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
      """
    +type: "link"
    +lang: "en"
    +isOc: false
    +hasEmbed: false
    +commentCount: 2
    +favouriteCount: 8
    +score: 0
    +isAdult: false
    +sticky: false
    +lastActive: DateTime @1688104448 {#1828
      date: 2023-06-30 07:54:08.0 +02:00
    }
    +ip: null
    +adaAmount: 0
    +tags: null
    +mentions: null
    +comments: Doctrine\ORM\PersistentCollection {#1905 …}
    +votes: Doctrine\ORM\PersistentCollection {#1934 …}
    +reports: Doctrine\ORM\PersistentCollection {#1929 …}
    +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
    +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
    +badges: Doctrine\ORM\PersistentCollection {#2357 …}
    +children: []
    -id: 547
    -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
    -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
    +cross: false
    +upVotes: 0
    +downVotes: 0
    +ranking: 1688088743
    +visibility: "visible             "
    +apId: "https://lemmy.world/post/804780"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688057743 {#2414
      date: 2023-06-29 18:55:43.0 +02:00
    }
  }
  +magazine: App\Entity\Magazine {#262}
  +image: null
  +parent: null
  +root: null
  +body: """
    As we are on the eve of rexxit - Is there a "best" way to sabotage our posts?\n
    \n
    I suppose I see two ways of achieving this - 1) a single AI-response that we edit all posts with; or 2) actually using an AI to "reply", as in different posts which emulate the answers a human would provide but generated by AI\n
    \n
    Imo, route 2 would be more time-consuming but harder to 'prevent' from degrading the dataset from reddit's perspective?
    """
  +lang: "en"
  +isAdult: false
  +favouriteCount: 1
  +score: 0
  +lastActive: DateTime @1688104452 {#4086
    date: 2023-06-30 07:54:12.0 +02:00
  }
  +ip: null
  +tags: null
  +mentions: [
    "@NevermindNoMind@lemmy.world"
  ]
  +children: Doctrine\ORM\PersistentCollection {#4075 …}
  +nested: Doctrine\ORM\PersistentCollection {#4073 …}
  +votes: Doctrine\ORM\PersistentCollection {#4071 …}
  +reports: Doctrine\ORM\PersistentCollection {#4069 …}
  +favourites: Doctrine\ORM\PersistentCollection {#4037 …}
  +notifications: Doctrine\ORM\PersistentCollection {#4041 …}
  -id: 5540
  -bodyTs: "'1':27 '2':40,65 'achiev':25 'actual':41 'ai':31,44,62 'ai-respons':30 'answer':54 'best':12 'consum':71 'dataset':79 'degrad':77 'differ':49 'edit':35 'emul':52 'eve':6 'generat':60 'harder':73 'human':56 'imo':63 'perspect':83 'post':17,37,50 'prevent':75 'provid':58 'reddit':81 'repli':46 'respons':32 'rexxit':8 'rout':64 'sabotag':15 'see':21 'singl':29 'suppos':19 'time':70 'time-consum':69 'two':22 'use':42 'way':13,23 'would':57,66"
  +ranking: 0
  +commentCount: 0
  +upVotes: 0
  +downVotes: 0
  +visibility: "visible             "
  +apId: "https://lemmy.world/comment/632155"
  +editedAt: null
  +createdAt: DateTimeImmutable @1688058152 {#4022
    date: 2023-06-29 19:02:32.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS DENIED
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
9 DENIED ROLE_USER
null
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"Symfony\Component\Security\Core\Authorization\Voter\RoleHierarchyVoter"
ACCESS DENIED
"App\Security\Voter\EntryCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
10 DENIED moderate
App\Entity\EntryComment {#4105
  +user: Proxies\__CG__\App\Entity\User {#4108 …}
  +entry: App\Entity\Entry {#2399
    +user: Proxies\__CG__\App\Entity\User {#1966 …}
    +magazine: App\Entity\Magazine {#262
      +icon: Proxies\__CG__\App\Entity\Image {#242 …}
      +name: "RedditMigration@kbin.social"
      +title: "RedditMigration"
      +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
      +rules: null
      +subscriptionsCount: 0
      +entryCount: 230
      +entryCommentCount: 3588
      +postCount: 46
      +postCommentCount: 100
      +isAdult: false
      +customCss: null
      +lastActive: DateTime @1689732875 {#270
        date: 2023-07-19 04:14:35.0 +02:00
      }
      +markedForDeletionAt: null
      +tags: null
      +moderators: Doctrine\ORM\PersistentCollection {#233 …}
      +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
      +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
      +entries: Doctrine\ORM\PersistentCollection {#176 …}
      +posts: Doctrine\ORM\PersistentCollection {#134 …}
      +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
      +bans: Doctrine\ORM\PersistentCollection {#113 …}
      +reports: Doctrine\ORM\PersistentCollection {#99 …}
      +badges: Doctrine\ORM\PersistentCollection {#77 …}
      +logs: Doctrine\ORM\PersistentCollection {#67 …}
      +awards: Doctrine\ORM\PersistentCollection {#2434 …}
      +categories: Doctrine\ORM\PersistentCollection {#1804 …}
      -id: 19
      +apId: "RedditMigration@kbin.social"
      +apProfileId: "https://kbin.social/m/RedditMigration"
      +apPublicUrl: "https://kbin.social/m/RedditMigration"
      +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
      +apInboxUrl: "https://kbin.social/f/inbox"
      +apDomain: "kbin.social"
      +apPreferredUsername: "RedditMigration"
      +apDiscoverable: true
      +apManuallyApprovesFollowers: null
      +privateKey: null
      +publicKey: null
      +apFetchedAt: DateTime @1688009400 {#268
        date: 2023-06-29 05:30:00.0 +02:00
      }
      +apDeletedAt: null
      +apTimeoutAt: null
      +visibility: "visible             "
      +createdAt: DateTimeImmutable @1688009399 {#276
        date: 2023-06-29 05:29:59.0 +02:00
      }
    }
    +image: null
    +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
    +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
    +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
    +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
    +body: """
      A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
      \n
      I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
      \n
      > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
      \n
      > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
      \n
      > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
      \n
      > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
      \n
      > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
      \n
      Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
      \n
      > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
      """
    +type: "link"
    +lang: "en"
    +isOc: false
    +hasEmbed: false
    +commentCount: 2
    +favouriteCount: 8
    +score: 0
    +isAdult: false
    +sticky: false
    +lastActive: DateTime @1688104448 {#1828
      date: 2023-06-30 07:54:08.0 +02:00
    }
    +ip: null
    +adaAmount: 0
    +tags: null
    +mentions: null
    +comments: Doctrine\ORM\PersistentCollection {#1905 …}
    +votes: Doctrine\ORM\PersistentCollection {#1934 …}
    +reports: Doctrine\ORM\PersistentCollection {#1929 …}
    +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
    +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
    +badges: Doctrine\ORM\PersistentCollection {#2357 …}
    +children: []
    -id: 547
    -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
    -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
    +cross: false
    +upVotes: 0
    +downVotes: 0
    +ranking: 1688088743
    +visibility: "visible             "
    +apId: "https://lemmy.world/post/804780"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688057743 {#2414
      date: 2023-06-29 18:55:43.0 +02:00
    }
  }
  +magazine: App\Entity\Magazine {#262}
  +image: null
  +parent: App\Entity\EntryComment {#4077
    +user: App\Entity\User {#4025 …}
    +entry: App\Entity\Entry {#2399}
    +magazine: App\Entity\Magazine {#262}
    +image: null
    +parent: null
    +root: null
    +body: """
      As we are on the eve of rexxit - Is there a "best" way to sabotage our posts?\n
      \n
      I suppose I see two ways of achieving this - 1) a single AI-response that we edit all posts with; or 2) actually using an AI to "reply", as in different posts which emulate the answers a human would provide but generated by AI\n
      \n
      Imo, route 2 would be more time-consuming but harder to 'prevent' from degrading the dataset from reddit's perspective?
      """
    +lang: "en"
    +isAdult: false
    +favouriteCount: 1
    +score: 0
    +lastActive: DateTime @1688104452 {#4086
      date: 2023-06-30 07:54:12.0 +02:00
    }
    +ip: null
    +tags: null
    +mentions: [
      "@NevermindNoMind@lemmy.world"
    ]
    +children: Doctrine\ORM\PersistentCollection {#4075 …}
    +nested: Doctrine\ORM\PersistentCollection {#4073 …}
    +votes: Doctrine\ORM\PersistentCollection {#4071 …}
    +reports: Doctrine\ORM\PersistentCollection {#4069 …}
    +favourites: Doctrine\ORM\PersistentCollection {#4037 …}
    +notifications: Doctrine\ORM\PersistentCollection {#4041 …}
    -id: 5540
    -bodyTs: "'1':27 '2':40,65 'achiev':25 'actual':41 'ai':31,44,62 'ai-respons':30 'answer':54 'best':12 'consum':71 'dataset':79 'degrad':77 'differ':49 'edit':35 'emul':52 'eve':6 'generat':60 'harder':73 'human':56 'imo':63 'perspect':83 'post':17,37,50 'prevent':75 'provid':58 'reddit':81 'repli':46 'respons':32 'rexxit':8 'rout':64 'sabotag':15 'see':21 'singl':29 'suppos':19 'time':70 'time-consum':69 'two':22 'use':42 'way':13,23 'would':57,66"
    +ranking: 0
    +commentCount: 0
    +upVotes: 0
    +downVotes: 0
    +visibility: "visible             "
    +apId: "https://lemmy.world/comment/632155"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688058152 {#4022
      date: 2023-06-29 19:02:32.0 +02:00
    }
  }
  +root: App\Entity\EntryComment {#4077}
  +body: """
    I used a free download called Redact to go through all my comments on June 11 and replace with AI language garbage. I did not delete submissions at this time, however, though that is an option in Redact. This process took almost 4 hours because I had two 11+ year old accounts.\n
    \n
    Because I started this late at night and am in a specific time zone, a few of the subs I commented in the most had gone dark (midnight of June 12) and my comments could not be edited on my SFW account. In doing this, I was permabanned from several subreddits on my NSFW account.\n
    \n
    Today, I opened Redact again to see if I could alter comments/remove submissions on my account that had the most subs go dark. Redact wouldn't even run for my SFW account so I logged in to reddit directly and saw a message that my account had been deactivated, which is why I think Redact was throwing me errors. I manually deleted all my submissions from both my accounts and manually deleted any comments that were original language from me.\n
    \n
    I left up the AI edited comments and then deleted both my accounts.
    """
  +lang: "en"
  +isAdult: false
  +favouriteCount: 0
  +score: 0
  +lastActive: DateTime @1688104448 {#4100
    date: 2023-06-30 07:54:08.0 +02:00
  }
  +ip: null
  +tags: null
  +mentions: [
    "@NevermindNoMind@lemmy.world"
    "@Methylman@lemmy.world"
  ]
  +children: Doctrine\ORM\PersistentCollection {#4104 …}
  +nested: Doctrine\ORM\PersistentCollection {#4113 …}
  +votes: Doctrine\ORM\PersistentCollection {#4094 …}
  +reports: Doctrine\ORM\PersistentCollection {#4092 …}
  +favourites: Doctrine\ORM\PersistentCollection {#4090 …}
  +notifications: Doctrine\ORM\PersistentCollection {#4088 …}
  -id: 5868
  -bodyTs: "'11':16,49 '12':83 '4':43 'account':52,94,107,123,139,153,176,200 'ai':20,192 'almost':42 'alter':118 'call':6 'comment':13,73,86,181,194 'comments/remove':119 'could':87,117 'dark':79,130 'deactiv':156 'delet':26,169,179,197 'direct':146 'download':5 'edit':90,193 'error':166 'even':134 'free':4 'garbag':22 'go':9,129 'gone':78 'hour':44 'howev':31 'june':15,82 'languag':21,185 'late':57 'left':189 'log':142 'manual':168,178 'messag':150 'midnight':80 'night':59 'nsfw':106 'old':51 'open':110 'option':36 'origin':184 'permaban':100 'process':40 'redact':7,38,111,131,162 'reddit':145 'replac':18 'run':135 'saw':148 'see':114 'sever':102 'sfw':93,138 'specif':64 'start':55 'sub':71,128 'submiss':27,120,172 'subreddit':103 'think':161 'though':32 'throw':164 'time':30,65 'today':108 'took':41 'two':48 'use':2 'wouldn':132 'year':50 'zone':66"
  +ranking: 0
  +commentCount: 0
  +upVotes: 0
  +downVotes: 0
  +visibility: "visible             "
  +apId: "https://kbin.social/m/RedditMigration/t/113497/-/comment/458083"
  +editedAt: DateTimeImmutable @1688104624 {#4097
    date: 2023-06-30 07:57:04.0 +02:00
  }
  +createdAt: DateTimeImmutable @1688104448 {#4095
    date: 2023-06-30 07:54:08.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS DENIED
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
11 DENIED edit
App\Entity\EntryComment {#4105
  +user: Proxies\__CG__\App\Entity\User {#4108 …}
  +entry: App\Entity\Entry {#2399
    +user: Proxies\__CG__\App\Entity\User {#1966 …}
    +magazine: App\Entity\Magazine {#262
      +icon: Proxies\__CG__\App\Entity\Image {#242 …}
      +name: "RedditMigration@kbin.social"
      +title: "RedditMigration"
      +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
      +rules: null
      +subscriptionsCount: 0
      +entryCount: 230
      +entryCommentCount: 3588
      +postCount: 46
      +postCommentCount: 100
      +isAdult: false
      +customCss: null
      +lastActive: DateTime @1689732875 {#270
        date: 2023-07-19 04:14:35.0 +02:00
      }
      +markedForDeletionAt: null
      +tags: null
      +moderators: Doctrine\ORM\PersistentCollection {#233 …}
      +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
      +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
      +entries: Doctrine\ORM\PersistentCollection {#176 …}
      +posts: Doctrine\ORM\PersistentCollection {#134 …}
      +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
      +bans: Doctrine\ORM\PersistentCollection {#113 …}
      +reports: Doctrine\ORM\PersistentCollection {#99 …}
      +badges: Doctrine\ORM\PersistentCollection {#77 …}
      +logs: Doctrine\ORM\PersistentCollection {#67 …}
      +awards: Doctrine\ORM\PersistentCollection {#2434 …}
      +categories: Doctrine\ORM\PersistentCollection {#1804 …}
      -id: 19
      +apId: "RedditMigration@kbin.social"
      +apProfileId: "https://kbin.social/m/RedditMigration"
      +apPublicUrl: "https://kbin.social/m/RedditMigration"
      +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
      +apInboxUrl: "https://kbin.social/f/inbox"
      +apDomain: "kbin.social"
      +apPreferredUsername: "RedditMigration"
      +apDiscoverable: true
      +apManuallyApprovesFollowers: null
      +privateKey: null
      +publicKey: null
      +apFetchedAt: DateTime @1688009400 {#268
        date: 2023-06-29 05:30:00.0 +02:00
      }
      +apDeletedAt: null
      +apTimeoutAt: null
      +visibility: "visible             "
      +createdAt: DateTimeImmutable @1688009399 {#276
        date: 2023-06-29 05:29:59.0 +02:00
      }
    }
    +image: null
    +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
    +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
    +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
    +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
    +body: """
      A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
      \n
      I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
      \n
      > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
      \n
      > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
      \n
      > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
      \n
      > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
      \n
      > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
      \n
      Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
      \n
      > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
      """
    +type: "link"
    +lang: "en"
    +isOc: false
    +hasEmbed: false
    +commentCount: 2
    +favouriteCount: 8
    +score: 0
    +isAdult: false
    +sticky: false
    +lastActive: DateTime @1688104448 {#1828
      date: 2023-06-30 07:54:08.0 +02:00
    }
    +ip: null
    +adaAmount: 0
    +tags: null
    +mentions: null
    +comments: Doctrine\ORM\PersistentCollection {#1905 …}
    +votes: Doctrine\ORM\PersistentCollection {#1934 …}
    +reports: Doctrine\ORM\PersistentCollection {#1929 …}
    +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
    +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
    +badges: Doctrine\ORM\PersistentCollection {#2357 …}
    +children: []
    -id: 547
    -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
    -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
    +cross: false
    +upVotes: 0
    +downVotes: 0
    +ranking: 1688088743
    +visibility: "visible             "
    +apId: "https://lemmy.world/post/804780"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688057743 {#2414
      date: 2023-06-29 18:55:43.0 +02:00
    }
  }
  +magazine: App\Entity\Magazine {#262}
  +image: null
  +parent: App\Entity\EntryComment {#4077
    +user: App\Entity\User {#4025 …}
    +entry: App\Entity\Entry {#2399}
    +magazine: App\Entity\Magazine {#262}
    +image: null
    +parent: null
    +root: null
    +body: """
      As we are on the eve of rexxit - Is there a "best" way to sabotage our posts?\n
      \n
      I suppose I see two ways of achieving this - 1) a single AI-response that we edit all posts with; or 2) actually using an AI to "reply", as in different posts which emulate the answers a human would provide but generated by AI\n
      \n
      Imo, route 2 would be more time-consuming but harder to 'prevent' from degrading the dataset from reddit's perspective?
      """
    +lang: "en"
    +isAdult: false
    +favouriteCount: 1
    +score: 0
    +lastActive: DateTime @1688104452 {#4086
      date: 2023-06-30 07:54:12.0 +02:00
    }
    +ip: null
    +tags: null
    +mentions: [
      "@NevermindNoMind@lemmy.world"
    ]
    +children: Doctrine\ORM\PersistentCollection {#4075 …}
    +nested: Doctrine\ORM\PersistentCollection {#4073 …}
    +votes: Doctrine\ORM\PersistentCollection {#4071 …}
    +reports: Doctrine\ORM\PersistentCollection {#4069 …}
    +favourites: Doctrine\ORM\PersistentCollection {#4037 …}
    +notifications: Doctrine\ORM\PersistentCollection {#4041 …}
    -id: 5540
    -bodyTs: "'1':27 '2':40,65 'achiev':25 'actual':41 'ai':31,44,62 'ai-respons':30 'answer':54 'best':12 'consum':71 'dataset':79 'degrad':77 'differ':49 'edit':35 'emul':52 'eve':6 'generat':60 'harder':73 'human':56 'imo':63 'perspect':83 'post':17,37,50 'prevent':75 'provid':58 'reddit':81 'repli':46 'respons':32 'rexxit':8 'rout':64 'sabotag':15 'see':21 'singl':29 'suppos':19 'time':70 'time-consum':69 'two':22 'use':42 'way':13,23 'would':57,66"
    +ranking: 0
    +commentCount: 0
    +upVotes: 0
    +downVotes: 0
    +visibility: "visible             "
    +apId: "https://lemmy.world/comment/632155"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688058152 {#4022
      date: 2023-06-29 19:02:32.0 +02:00
    }
  }
  +root: App\Entity\EntryComment {#4077}
  +body: """
    I used a free download called Redact to go through all my comments on June 11 and replace with AI language garbage. I did not delete submissions at this time, however, though that is an option in Redact. This process took almost 4 hours because I had two 11+ year old accounts.\n
    \n
    Because I started this late at night and am in a specific time zone, a few of the subs I commented in the most had gone dark (midnight of June 12) and my comments could not be edited on my SFW account. In doing this, I was permabanned from several subreddits on my NSFW account.\n
    \n
    Today, I opened Redact again to see if I could alter comments/remove submissions on my account that had the most subs go dark. Redact wouldn't even run for my SFW account so I logged in to reddit directly and saw a message that my account had been deactivated, which is why I think Redact was throwing me errors. I manually deleted all my submissions from both my accounts and manually deleted any comments that were original language from me.\n
    \n
    I left up the AI edited comments and then deleted both my accounts.
    """
  +lang: "en"
  +isAdult: false
  +favouriteCount: 0
  +score: 0
  +lastActive: DateTime @1688104448 {#4100
    date: 2023-06-30 07:54:08.0 +02:00
  }
  +ip: null
  +tags: null
  +mentions: [
    "@NevermindNoMind@lemmy.world"
    "@Methylman@lemmy.world"
  ]
  +children: Doctrine\ORM\PersistentCollection {#4104 …}
  +nested: Doctrine\ORM\PersistentCollection {#4113 …}
  +votes: Doctrine\ORM\PersistentCollection {#4094 …}
  +reports: Doctrine\ORM\PersistentCollection {#4092 …}
  +favourites: Doctrine\ORM\PersistentCollection {#4090 …}
  +notifications: Doctrine\ORM\PersistentCollection {#4088 …}
  -id: 5868
  -bodyTs: "'11':16,49 '12':83 '4':43 'account':52,94,107,123,139,153,176,200 'ai':20,192 'almost':42 'alter':118 'call':6 'comment':13,73,86,181,194 'comments/remove':119 'could':87,117 'dark':79,130 'deactiv':156 'delet':26,169,179,197 'direct':146 'download':5 'edit':90,193 'error':166 'even':134 'free':4 'garbag':22 'go':9,129 'gone':78 'hour':44 'howev':31 'june':15,82 'languag':21,185 'late':57 'left':189 'log':142 'manual':168,178 'messag':150 'midnight':80 'night':59 'nsfw':106 'old':51 'open':110 'option':36 'origin':184 'permaban':100 'process':40 'redact':7,38,111,131,162 'reddit':145 'replac':18 'run':135 'saw':148 'see':114 'sever':102 'sfw':93,138 'specif':64 'start':55 'sub':71,128 'submiss':27,120,172 'subreddit':103 'think':161 'though':32 'throw':164 'time':30,65 'today':108 'took':41 'two':48 'use':2 'wouldn':132 'year':50 'zone':66"
  +ranking: 0
  +commentCount: 0
  +upVotes: 0
  +downVotes: 0
  +visibility: "visible             "
  +apId: "https://kbin.social/m/RedditMigration/t/113497/-/comment/458083"
  +editedAt: DateTimeImmutable @1688104624 {#4097
    date: 2023-06-30 07:57:04.0 +02:00
  }
  +createdAt: DateTimeImmutable @1688104448 {#4095
    date: 2023-06-30 07:54:08.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS DENIED
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
12 DENIED moderate
App\Entity\EntryComment {#4105
  +user: Proxies\__CG__\App\Entity\User {#4108 …}
  +entry: App\Entity\Entry {#2399
    +user: Proxies\__CG__\App\Entity\User {#1966 …}
    +magazine: App\Entity\Magazine {#262
      +icon: Proxies\__CG__\App\Entity\Image {#242 …}
      +name: "RedditMigration@kbin.social"
      +title: "RedditMigration"
      +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
      +rules: null
      +subscriptionsCount: 0
      +entryCount: 230
      +entryCommentCount: 3588
      +postCount: 46
      +postCommentCount: 100
      +isAdult: false
      +customCss: null
      +lastActive: DateTime @1689732875 {#270
        date: 2023-07-19 04:14:35.0 +02:00
      }
      +markedForDeletionAt: null
      +tags: null
      +moderators: Doctrine\ORM\PersistentCollection {#233 …}
      +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
      +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
      +entries: Doctrine\ORM\PersistentCollection {#176 …}
      +posts: Doctrine\ORM\PersistentCollection {#134 …}
      +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
      +bans: Doctrine\ORM\PersistentCollection {#113 …}
      +reports: Doctrine\ORM\PersistentCollection {#99 …}
      +badges: Doctrine\ORM\PersistentCollection {#77 …}
      +logs: Doctrine\ORM\PersistentCollection {#67 …}
      +awards: Doctrine\ORM\PersistentCollection {#2434 …}
      +categories: Doctrine\ORM\PersistentCollection {#1804 …}
      -id: 19
      +apId: "RedditMigration@kbin.social"
      +apProfileId: "https://kbin.social/m/RedditMigration"
      +apPublicUrl: "https://kbin.social/m/RedditMigration"
      +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
      +apInboxUrl: "https://kbin.social/f/inbox"
      +apDomain: "kbin.social"
      +apPreferredUsername: "RedditMigration"
      +apDiscoverable: true
      +apManuallyApprovesFollowers: null
      +privateKey: null
      +publicKey: null
      +apFetchedAt: DateTime @1688009400 {#268
        date: 2023-06-29 05:30:00.0 +02:00
      }
      +apDeletedAt: null
      +apTimeoutAt: null
      +visibility: "visible             "
      +createdAt: DateTimeImmutable @1688009399 {#276
        date: 2023-06-29 05:29:59.0 +02:00
      }
    }
    +image: null
    +domain: Proxies\__CG__\App\Entity\Domain {#1900 …}
    +slug: "Overwriting-Comments-w-AI-Output-Is-the-Quickest-Way-to"
    +title: "Overwriting Comments w/ AI Output Is the Quickest Way to Make Reddit's Data Useless to LLM Firms"
    +url: "https://arxiv.org/pdf/2305.17493v2.pdf"
    +body: """
      A new study shows that LLM models that are fed too much content that was generated by LLMs eventually collapse. Essentially, text generated by AI is poison if it makes its way into an LLMs training data. If the model eats too much of this poison, the model dies. By replacing your Reddit comments with AI generated text, you can effectively increase the toxicity of Reddit's dataset, and thereby decrease its value to firms training new LLMs. This will probably happen naturally anyway as spam bots and so forth continue taking over Reddit, but if you want to go out in a petty way, this is a good option.\n
      \n
      I linked the actual study, but I first read about this on [Platformer](https://www.platformer.news/p/the-ai-is-eating-itself), where he was writing more broadly about how the AI is filing up the web with synthetic content and the problems that is causing. He was using this study to point out that it will be increasingly hard for developers to find good content for the LLMs to train on due to there being so much AI generated content, and the risk of the LLMs consuming too much AI content. Here is what he wrote:\n
      \n
      > A second, more worrisome study comes from researchers at the University of Oxford, University of Cambridge, University of Toronto, and Imperial College London. It found that training AI systems on data generated by other AI systems — synthetic data, to use the industry’s term — causes models to degrade and ultimately collapse.\n
      \n
      > While the decay can be managed by using synthetic data sparingly, researchers write, the idea that models can be “poisoned” by feeding them their own outputs raises real risks for the web.\n
      \n
      > And that’s a problem, because — to bring together the threads of today’s newsletter so far — AI output is spreading to encompass more of the web every day.\n
      \n
      > “The obvious larger question,” Clark writes, “is what this does to competition among AI developers as the internet fills up with a greater percentage of generated versus real content.”\n
      \n
      > When tech companies were building the first chatbots, they could be certain that the vast majority of the data they were scraping was human-generated. Going forward, though, they’ll be ever less certain of that — and until they figure out reliable ways to identify chatbot-generated text, they’re at risk of breaking their own models.\n
      \n
      Even the study's abstract doesn't make a lot of sense to me, so here is an AI generated ELI5 (I am fully aware of the irony):\n
      \n
      > This paper is about how computers learn to write like humans. They use a lot of text from the internet to learn how to write. But if they use too much text that they wrote themselves, they start to forget how humans write. This is bad because we want computers to write like humans. So we need to make sure that computers learn from humans and not just from other computers.
      """
    +type: "link"
    +lang: "en"
    +isOc: false
    +hasEmbed: false
    +commentCount: 2
    +favouriteCount: 8
    +score: 0
    +isAdult: false
    +sticky: false
    +lastActive: DateTime @1688104448 {#1828
      date: 2023-06-30 07:54:08.0 +02:00
    }
    +ip: null
    +adaAmount: 0
    +tags: null
    +mentions: null
    +comments: Doctrine\ORM\PersistentCollection {#1905 …}
    +votes: Doctrine\ORM\PersistentCollection {#1934 …}
    +reports: Doctrine\ORM\PersistentCollection {#1929 …}
    +favourites: Doctrine\ORM\PersistentCollection {#1396 …}
    +notifications: Doctrine\ORM\PersistentCollection {#2439 …}
    +badges: Doctrine\ORM\PersistentCollection {#2357 …}
    +children: []
    -id: 547
    -titleTs: "'ai':4 'comment':2 'data':14 'firm':18 'llm':17 'make':11 'output':5 'overwrit':1 'quickest':8 'reddit':12 'useless':15 'w':3 'way':9"
    -bodyTs: "'/p/the-ai-is-eating-itself),':126 'abstract':406 'actual':114 'ai':25,56,136,183,195,229,236,302,327,420 'among':326 'anyway':84 'awar':426 'bad':475 'bot':87 'break':398 'bring':292 'broad':132 'build':347 'cambridg':217 'caus':150,246 'certain':354,377 'chatbot':350,390 'chatbot-gener':389 'clark':318 'collaps':20,252 'colleg':223 'come':207 'comment':54 'compani':345 'competit':325 'comput':435,479,491,500 'consum':192 'content':13,144,170,185,196,342 'continu':91 'could':352 'data':37,232,239,262,361 'dataset':68 'day':313 'decay':255 'decreas':71 'degrad':249 'develop':166,328 'die':49 'doesn':407 'due':177 'eat':41 'effect':61 'eli5':422 'encompass':307 'essenti':21 'even':402 'eventu':19 'ever':375 'everi':312 'far':301 'fed':10 'feed':274 'figur':383 'file':138 'fill':332 'find':168 'firm':75 'first':118,349 'forget':469 'forth':90 'forward':370 'found':226 'fulli':425 'generat':16,23,57,184,233,339,368,391,421 'go':100,369 'good':109,169 'greater':336 'happen':82 'hard':164 'human':367,440,471,483,494 'human-gener':366 'idea':267 'identifi':388 'imperi':222 'increas':62,163 'industri':243 'internet':331,449 'ironi':429 'larger':316 'learn':436,451,492 'less':376 'like':439,482 'link':112 'll':373 'llm':6 'llms':18,35,78,173,191 'london':224 'lot':411,444 'major':358 'make':30,409,488 'manag':258 'model':7,40,48,247,269,401 'much':12,43,182,194,460 'natur':83 'need':486 'new':2,77 'newslett':299 'obvious':315 'option':110 'output':278,303 'oxford':214 'paper':431 'percentag':337 'petti':104 'platform':123 'point':157 'poison':27,46,272 'probabl':81 'problem':147,289 'question':317 'rais':279 're':394 'read':119 'real':280,341 'reddit':53,66,94 'reliabl':385 'replac':51 'research':209,264 'risk':188,281,396 'scrape':364 'second':203 'sens':413 'show':4 'spam':86 'spare':263 'spread':305 'start':467 'studi':3,115,155,206,404 'sure':489 'synthet':143,238,261 'system':230,237 'take':92 'tech':344 'term':245 'text':22,58,392,446,461 'therebi':70 'though':371 'thread':295 'today':297 'togeth':293 'toronto':220 'toxic':64 'train':36,76,175,228 'ultim':251 'univers':212,215,218 'use':153,241,260,442,458 'valu':73 'vast':357 'versus':340 'want':98,478 'way':32,105,386 'web':141,284,311 'worrisom':205 'write':130,265,319,438,454,472,481 'wrote':201,464 'www.platformer.news':125 'www.platformer.news/p/the-ai-is-eating-itself),':124"
    +cross: false
    +upVotes: 0
    +downVotes: 0
    +ranking: 1688088743
    +visibility: "visible             "
    +apId: "https://lemmy.world/post/804780"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688057743 {#2414
      date: 2023-06-29 18:55:43.0 +02:00
    }
  }
  +magazine: App\Entity\Magazine {#262}
  +image: null
  +parent: App\Entity\EntryComment {#4077
    +user: App\Entity\User {#4025 …}
    +entry: App\Entity\Entry {#2399}
    +magazine: App\Entity\Magazine {#262}
    +image: null
    +parent: null
    +root: null
    +body: """
      As we are on the eve of rexxit - Is there a "best" way to sabotage our posts?\n
      \n
      I suppose I see two ways of achieving this - 1) a single AI-response that we edit all posts with; or 2) actually using an AI to "reply", as in different posts which emulate the answers a human would provide but generated by AI\n
      \n
      Imo, route 2 would be more time-consuming but harder to 'prevent' from degrading the dataset from reddit's perspective?
      """
    +lang: "en"
    +isAdult: false
    +favouriteCount: 1
    +score: 0
    +lastActive: DateTime @1688104452 {#4086
      date: 2023-06-30 07:54:12.0 +02:00
    }
    +ip: null
    +tags: null
    +mentions: [
      "@NevermindNoMind@lemmy.world"
    ]
    +children: Doctrine\ORM\PersistentCollection {#4075 …}
    +nested: Doctrine\ORM\PersistentCollection {#4073 …}
    +votes: Doctrine\ORM\PersistentCollection {#4071 …}
    +reports: Doctrine\ORM\PersistentCollection {#4069 …}
    +favourites: Doctrine\ORM\PersistentCollection {#4037 …}
    +notifications: Doctrine\ORM\PersistentCollection {#4041 …}
    -id: 5540
    -bodyTs: "'1':27 '2':40,65 'achiev':25 'actual':41 'ai':31,44,62 'ai-respons':30 'answer':54 'best':12 'consum':71 'dataset':79 'degrad':77 'differ':49 'edit':35 'emul':52 'eve':6 'generat':60 'harder':73 'human':56 'imo':63 'perspect':83 'post':17,37,50 'prevent':75 'provid':58 'reddit':81 'repli':46 'respons':32 'rexxit':8 'rout':64 'sabotag':15 'see':21 'singl':29 'suppos':19 'time':70 'time-consum':69 'two':22 'use':42 'way':13,23 'would':57,66"
    +ranking: 0
    +commentCount: 0
    +upVotes: 0
    +downVotes: 0
    +visibility: "visible             "
    +apId: "https://lemmy.world/comment/632155"
    +editedAt: null
    +createdAt: DateTimeImmutable @1688058152 {#4022
      date: 2023-06-29 19:02:32.0 +02:00
    }
  }
  +root: App\Entity\EntryComment {#4077}
  +body: """
    I used a free download called Redact to go through all my comments on June 11 and replace with AI language garbage. I did not delete submissions at this time, however, though that is an option in Redact. This process took almost 4 hours because I had two 11+ year old accounts.\n
    \n
    Because I started this late at night and am in a specific time zone, a few of the subs I commented in the most had gone dark (midnight of June 12) and my comments could not be edited on my SFW account. In doing this, I was permabanned from several subreddits on my NSFW account.\n
    \n
    Today, I opened Redact again to see if I could alter comments/remove submissions on my account that had the most subs go dark. Redact wouldn't even run for my SFW account so I logged in to reddit directly and saw a message that my account had been deactivated, which is why I think Redact was throwing me errors. I manually deleted all my submissions from both my accounts and manually deleted any comments that were original language from me.\n
    \n
    I left up the AI edited comments and then deleted both my accounts.
    """
  +lang: "en"
  +isAdult: false
  +favouriteCount: 0
  +score: 0
  +lastActive: DateTime @1688104448 {#4100
    date: 2023-06-30 07:54:08.0 +02:00
  }
  +ip: null
  +tags: null
  +mentions: [
    "@NevermindNoMind@lemmy.world"
    "@Methylman@lemmy.world"
  ]
  +children: Doctrine\ORM\PersistentCollection {#4104 …}
  +nested: Doctrine\ORM\PersistentCollection {#4113 …}
  +votes: Doctrine\ORM\PersistentCollection {#4094 …}
  +reports: Doctrine\ORM\PersistentCollection {#4092 …}
  +favourites: Doctrine\ORM\PersistentCollection {#4090 …}
  +notifications: Doctrine\ORM\PersistentCollection {#4088 …}
  -id: 5868
  -bodyTs: "'11':16,49 '12':83 '4':43 'account':52,94,107,123,139,153,176,200 'ai':20,192 'almost':42 'alter':118 'call':6 'comment':13,73,86,181,194 'comments/remove':119 'could':87,117 'dark':79,130 'deactiv':156 'delet':26,169,179,197 'direct':146 'download':5 'edit':90,193 'error':166 'even':134 'free':4 'garbag':22 'go':9,129 'gone':78 'hour':44 'howev':31 'june':15,82 'languag':21,185 'late':57 'left':189 'log':142 'manual':168,178 'messag':150 'midnight':80 'night':59 'nsfw':106 'old':51 'open':110 'option':36 'origin':184 'permaban':100 'process':40 'redact':7,38,111,131,162 'reddit':145 'replac':18 'run':135 'saw':148 'see':114 'sever':102 'sfw':93,138 'specif':64 'start':55 'sub':71,128 'submiss':27,120,172 'subreddit':103 'think':161 'though':32 'throw':164 'time':30,65 'today':108 'took':41 'two':48 'use':2 'wouldn':132 'year':50 'zone':66"
  +ranking: 0
  +commentCount: 0
  +upVotes: 0
  +downVotes: 0
  +visibility: "visible             "
  +apId: "https://kbin.social/m/RedditMigration/t/113497/-/comment/458083"
  +editedAt: DateTimeImmutable @1688104624 {#4097
    date: 2023-06-30 07:57:04.0 +02:00
  }
  +createdAt: DateTimeImmutable @1688104448 {#4095
    date: 2023-06-30 07:54:08.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS DENIED
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details
13 DENIED edit
App\Entity\Magazine {#262
  +icon: Proxies\__CG__\App\Entity\Image {#242 …}
  +name: "RedditMigration@kbin.social"
  +title: "RedditMigration"
  +description: "Tracking the lastest news and numbers about the #RedditMigration to open, Fediverse-based alternatives, including #Kbin and #Lemmy To see latest reeddit blackout info, see here: https://reddark.untone.uk/"
  +rules: null
  +subscriptionsCount: 0
  +entryCount: 230
  +entryCommentCount: 3588
  +postCount: 46
  +postCommentCount: 100
  +isAdult: false
  +customCss: null
  +lastActive: DateTime @1689732875 {#270
    date: 2023-07-19 04:14:35.0 +02:00
  }
  +markedForDeletionAt: null
  +tags: null
  +moderators: Doctrine\ORM\PersistentCollection {#233 …}
  +ownershipRequests: Doctrine\ORM\PersistentCollection {#229 …}
  +moderatorRequests: Doctrine\ORM\PersistentCollection {#218 …}
  +entries: Doctrine\ORM\PersistentCollection {#176 …}
  +posts: Doctrine\ORM\PersistentCollection {#134 …}
  +subscriptions: Doctrine\ORM\PersistentCollection {#196 …}
  +bans: Doctrine\ORM\PersistentCollection {#113 …}
  +reports: Doctrine\ORM\PersistentCollection {#99 …}
  +badges: Doctrine\ORM\PersistentCollection {#77 …}
  +logs: Doctrine\ORM\PersistentCollection {#67 …}
  +awards: Doctrine\ORM\PersistentCollection {#2434 …}
  +categories: Doctrine\ORM\PersistentCollection {#1804 …}
  -id: 19
  +apId: "RedditMigration@kbin.social"
  +apProfileId: "https://kbin.social/m/RedditMigration"
  +apPublicUrl: "https://kbin.social/m/RedditMigration"
  +apFollowersUrl: "https://kbin.social/m/RedditMigration/followers"
  +apInboxUrl: "https://kbin.social/f/inbox"
  +apDomain: "kbin.social"
  +apPreferredUsername: "RedditMigration"
  +apDiscoverable: true
  +apManuallyApprovesFollowers: null
  +privateKey: null
  +publicKey: null
  +apFetchedAt: DateTime @1688009400 {#268
    date: 2023-06-29 05:30:00.0 +02:00
  }
  +apDeletedAt: null
  +apTimeoutAt: null
  +visibility: "visible             "
  +createdAt: DateTimeImmutable @1688009399 {#276
    date: 2023-06-29 05:29:59.0 +02:00
  }
}
"Scheb\TwoFactorBundle\Security\Authorization\Voter\TwoFactorInProgressVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\EntryVoter"
ACCESS ABSTAIN
"App\Security\Voter\MagazineVoter"
ACCESS DENIED
"App\Security\Voter\MessageThreadVoter"
ACCESS ABSTAIN
"App\Security\Voter\MessageVoter"
ACCESS ABSTAIN
"App\Security\Voter\NotificationVoter"
ACCESS ABSTAIN
"App\Security\Voter\OAuth2UserConsentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostCommentVoter"
ACCESS ABSTAIN
"App\Security\Voter\PostVoter"
ACCESS ABSTAIN
"App\Security\Voter\UserVoter"
ACCESS ABSTAIN
Show voter details