SharedInboxController
Request
GET Parameters
None
POST Parameters
None
Uploaded Files
None
Request Attributes
Key | Value |
---|---|
_controller | "App\Controller\ActivityPub\SharedInboxController" |
_firewall_context | "security.firewall.map.context.main" |
_route | "ap_shared_inbox" |
_route_params | [] |
_security_firewall_run | "_security_main" |
_stopwatch_token | "55e9da" |
Request Headers
Header | Value |
---|---|
accept | "*/*" |
accept-encoding | "gzip" |
content-length | "9531" |
content-type | "application/activity+json" |
date | "Wed, 28 May 2025 20:56:17 GMT" |
digest | "SHA-256=cSiBieXOkr0ygkVhKE9Nr9xgXhY+r3yoW8O6fuH9ZV4=" |
host | "kbin.spritesserver.nl" |
signature | "keyId="https://lemmy.world/c/selfhosted#main-key",algorithm="hs2019",headers="(request-target) content-type date digest host",signature="FklvhoDl/5vnM0fzslD+hvd93MwpjjKGDDVZwsKiVuTuaKOk7eRqmAHaQKbY2oeocNvd3WTEuVhgNZBDs6tK5u2oMShl0caAmqgxh7yL90oAP73U2jVbArO7nOpk4/NRIpK41abPhnMFUIT/73SnpvoIGLhhPZXgwXtBn3ZdWa6DJt/4FHTvvZ2CDNfbilANFPn29yMPfJJ7Av7bhJcYzLVw2xc9yD5rcqtRwTpNcREcgnHkiT5Ut63cOaS8zxzmKDRZTwGKf3dkPTbi3Lzo56xkFfW1oXUUXMhjIAYSC3NzLClJyS1cHqBmioGIAWXtllR5dzcGl9csc6fi+/LS3g=="" |
traceparent | "00-e89f33937cf52c0e58d9770c68da3405-b599710561ead89a-01" |
tracestate | "" |
user-agent | "Lemmy/0.19.11-19-g2895f45e8; +https://lemmy.world" |
x-php-ob-level | "1" |
Request Content
Pretty
{ "@context": [ "https:\/\/join-lemmy.org\/context.json", "https:\/\/www.w3.org\/ns\/activitystreams" ], "actor": "https:\/\/lemmy.world\/c\/selfhosted", "to": [ "https:\/\/www.w3.org\/ns\/activitystreams#Public" ], "object": { "id": "https:\/\/lemmy.ca\/activities\/update\/ac8ea8ad-5dbd-4406-b922-113f661d0b82", "actor": "https:\/\/lemmy.ca\/u\/avidamoeba", "@context": [ "https:\/\/join-lemmy.org\/context.json", "https:\/\/www.w3.org\/ns\/activitystreams" ], "to": [ "https:\/\/www.w3.org\/ns\/activitystreams#Public" ], "object": { "type": "Note", "id": "https:\/\/lemmy.ca\/comment\/16817661", "attributedTo": "https:\/\/lemmy.ca\/u\/avidamoeba", "to": [ "https:\/\/www.w3.org\/ns\/activitystreams#Public" ], "cc": [ "https:\/\/lemmy.world\/c\/selfhosted", "https:\/\/lemmy.world\/u\/ZeDoTelhado" ], "content": "<p>That\u2019s really weird. I\u2019ve been using it for mobile-desktop-server-offsite sync for many years, with transfer sizes over 15TB, over WiFi, cellular, cable, fiber. I\u2019ve never seen data corruption. Conflicts, sometimes. Permission issues, sometimes. Wiping something accidentally, sometimes. It\u2019s even more weird because Syncthing performs computes hash values for the files it manages. I don\u2019t know if it performs hash validation after copying remotely but if not, it can be forced manually which would tell you what\u2019s fucked and be pulled from the source, if it still exists.<\/p>\n<p>Nevermind, <a href=\"https:\/\/docs.syncthing.net\/users\/syncing.html\" rel=\"nofollow\">it verifies the result:<\/a><\/p>\n<blockquote>\n<p>When a block is copied or received from another device, its SHA256 hash is computed and compared with the expected value. If it matches the block is written to a temporary copy of the file, otherwise it is discarded and Syncthing tries to find another source for the block.<\/p>\n<\/blockquote>\n<p>According to this, if you have data corruption it can only occur between copying\/moving a temporary file on your destination to another directory, or it could occur on the source itself. Both of those scenarios are a cause of concern and would likely persist with any utility. Moving or copying a file from one location to another on a sane machine should not corrupt it. If I were you I\u2019d ensure my server doesn\u2019t eat bits. If not the storage media, it could be bit rot, or bad RAM.<\/p>\n<p>Just in case everything seems fine, let me tell you what I dealt with. I had a Ryzen 5950X machine with 32GB of RAM. It worked well since inception with no signs of RAM or data corruption issues. I test every new machine with Memtest86+. At some point I migrated the storage from Ext4 on LVMRAID to ZFS. All good. Then I wrote an alarm for Prometheus to tell me if there\u2019s any issues in ZFS. A week later I get an email about a ZFS error. I check the system - says checksum errors, data has been corrected, applications unaffected, run a scrub to clear. I ran a scrub. A few more checksum errors found, all corrected, we\u2019re clean now. There was a strong solar storm around that time, probably that. A couple of weeks later I get another email. Same symptoms, same procedure. No solar storm. Shit. Memtest86+ - pass. Hm. A couple of weeks later I get another. Same thing. Memtest again - nothing. This went on for several months. Meanwhile the off-site backup sees nothing like that. While running Memtest on another machine I noticed that the test passes following the first took longer than the first, a lot longer. I thought something might be wrong with that machine. Dug into it, got into Memtest\u2019s source code and discovered that the first pass is shorter on purpose so that it quickly flags obviously bad RAM. Apparently if you want to detect less obvious issues, you have to run multiple passes. OK. Memtest the main server again, pass 1: OK, pass 2: OK, pass 3: OK, pass 4: FAIL. FUCK. Memtest each stick separately for 4 passes: OK. Memtest 2 at a time: OK. Memtest all 4: FAIL. Alright, now we know why ZFS keeps finding checksum errors. Long story short, this machine could not run this RAM in 4-DIMM config. Replaced it with another RAM that\u2019s rated to run in 4-DIMM config on that processor. No more checksum issues. If I was running the older Ext4-on-LVMRAID storage stack, I would have caught NONE of these and it would have happily corrupted files here and there. In fact it likely did and I have some corruption. Moral of the story - run <em>many<\/em> Memtest passes and use <em>checksumming<\/em> storage stack like ZFS or Btrfs. I strongly recommend ZFS since its stripe RAID works fine unlike Btrfs\u2019es. Start today, even if you\u2019re working with a single disk and add redundancy when you can. Only after change Syncthing for something else if you still somehow get corruption without ZFS\u2019es knowledge. And if ZFS tells you that you have checksum errors, you likely have bad hardware.<\/p>\n", "inReplyTo": "https:\/\/lemmy.world\/post\/30402224", "mediaType": "text\/html", "source": { "content": "That's really weird. I've been using it for mobile-desktop-server-offsite sync for many years, with transfer sizes over 15TB, over WiFi, cellular, cable, fiber. I've never seen data corruption. Conflicts, sometimes. Permission issues, sometimes. Wiping something accidentally, sometimes. It's even more weird because Syncthing performs computes hash values for the files it manages. I don't know if it performs hash validation after copying remotely but if not, it can be forced manually which would tell you what's fucked and be pulled from the source, if it still exists.\n\nNevermind, [it verifies the result:](https:\/\/docs.syncthing.net\/users\/syncing.html)\n\n> When a block is copied or received from another device, its SHA256 hash is computed and compared with the expected value. If it matches the block is written to a temporary copy of the file, otherwise it is discarded and Syncthing tries to find another source for the block.\n\nAccording to this, if you have data corruption it can only occur between copying\/moving a temporary file on your destination to another directory, or it could occur on the source itself. Both of those scenarios are a cause of concern and would likely persist with any utility. Moving or copying a file from one location to another on a sane machine should not corrupt it. If I were you I'd ensure my server doesn't eat bits. If not the storage media, it could be bit rot, or bad RAM.\n\nJust in case everything seems fine, let me tell you what I dealt with. I had a Ryzen 5950X machine with 32GB of RAM. It worked well since inception with no signs of RAM or data corruption issues. I test every new machine with Memtest86+. At some point I migrated the storage from Ext4 on LVMRAID to ZFS. All good. Then I wrote an alarm for Prometheus to tell me if there's any issues in ZFS. A week later I get an email about a ZFS error. I check the system - says checksum errors, data has been corrected, applications unaffected, run a scrub to clear. I ran a scrub. A few more checksum errors found, all corrected, we're clean now. There was a strong solar storm around that time, probably that. A couple of weeks later I get another email. Same symptoms, same procedure. No solar storm. Shit. Memtest86+ - pass. Hm. A couple of weeks later I get another. Same thing. Memtest again - nothing. This went on for several months. Meanwhile the off-site backup sees nothing like that. While running Memtest on another machine I noticed that the test passes following the first took longer than the first, a lot longer. I thought something might be wrong with that machine. Dug into it, got into Memtest's source code and discovered that the first pass is shorter on purpose so that it quickly flags obviously bad RAM. Apparently if you want to detect less obvious issues, you have to run multiple passes. OK. Memtest the main server again, pass 1: OK, pass 2: OK, pass 3: OK, pass 4: FAIL. FUCK. Memtest each stick separately for 4 passes: OK. Memtest 2 at a time: OK. Memtest all 4: FAIL. Alright, now we know why ZFS keeps finding checksum errors. Long story short, this machine could not run this RAM in 4-DIMM config. Replaced it with another RAM that's rated to run in 4-DIMM config on that processor. No more checksum issues. If I was running the older Ext4-on-LVMRAID storage stack, I would have caught NONE of these and it would have happily corrupted files here and there. In fact it likely did and I have some corruption. Moral of the story - run *many* Memtest passes and use *checksumming* storage stack like ZFS or Btrfs. I strongly recommend ZFS since its stripe RAID works fine unlike Btrfs'es. Start today, even if you're working with a single disk and add redundancy when you can. Only after change Syncthing for something else if you still somehow get corruption without ZFS'es knowledge. And if ZFS tells you that you have checksum errors, you likely have bad hardware.", "mediaType": "text\/markdown" }, "published": "2025-05-28T20:29:39.335302Z", "updated": "2025-05-28T20:56:05.287078Z", "tag": [ { "href": "https:\/\/lemmy.world\/u\/ZeDoTelhado", "name": "@ZeDoTelhado@lemmy.world", "type": "Mention" } ], "distinguished": false, "language": { "identifier": "en", "name": "English" }, "audience": "https:\/\/lemmy.world\/c\/selfhosted", "attachment": [] }, "cc": [ "https:\/\/lemmy.world\/c\/selfhosted", "https:\/\/lemmy.world\/u\/ZeDoTelhado" ], "tag": [ { "href": "https:\/\/lemmy.world\/u\/ZeDoTelhado", "name": "@ZeDoTelhado@lemmy.world", "type": "Mention" } ], "type": "Update", "audience": "https:\/\/lemmy.world\/c\/selfhosted" }, "cc": [ "https:\/\/lemmy.world\/c\/selfhosted\/followers" ], "type": "Announce", "id": "https:\/\/lemmy.world\/activities\/announce\/update\/3e19197c-7f1d-4c5d-a58f-338cb39261a8" }
Raw
{"@context":["https://join-lemmy.org/context.json","https://www.w3.org/ns/activitystreams"],"actor":"https://lemmy.world/c/selfhosted","to":["https://www.w3.org/ns/activitystreams#Public"],"object":{"id":"https://lemmy.ca/activities/update/ac8ea8ad-5dbd-4406-b922-113f661d0b82","actor":"https://lemmy.ca/u/avidamoeba","@context":["https://join-lemmy.org/context.json","https://www.w3.org/ns/activitystreams"],"to":["https://www.w3.org/ns/activitystreams#Public"],"object":{"type":"Note","id":"https://lemmy.ca/comment/16817661","attributedTo":"https://lemmy.ca/u/avidamoeba","to":["https://www.w3.org/ns/activitystreams#Public"],"cc":["https://lemmy.world/c/selfhosted","https://lemmy.world/u/ZeDoTelhado"],"content":"<p>That’s really weird. I’ve been using it for mobile-desktop-server-offsite sync for many years, with transfer sizes over 15TB, over WiFi, cellular, cable, fiber. I’ve never seen data corruption. Conflicts, sometimes. Permission issues, sometimes. Wiping something accidentally, sometimes. It’s even more weird because Syncthing performs computes hash values for the files it manages. I don’t know if it performs hash validation after copying remotely but if not, it can be forced manually which would tell you what’s fucked and be pulled from the source, if it still exists.</p>\n<p>Nevermind, <a href=\"https://docs.syncthing.net/users/syncing.html\" rel=\"nofollow\">it verifies the result:</a></p>\n<blockquote>\n<p>When a block is copied or received from another device, its SHA256 hash is computed and compared with the expected value. If it matches the block is written to a temporary copy of the file, otherwise it is discarded and Syncthing tries to find another source for the block.</p>\n</blockquote>\n<p>According to this, if you have data corruption it can only occur between copying/moving a temporary file on your destination to another directory, or it could occur on the source itself. Both of those scenarios are a cause of concern and would likely persist with any utility. Moving or copying a file from one location to another on a sane machine should not corrupt it. If I were you I’d ensure my server doesn’t eat bits. If not the storage media, it could be bit rot, or bad RAM.</p>\n<p>Just in case everything seems fine, let me tell you what I dealt with. I had a Ryzen 5950X machine with 32GB of RAM. It worked well since inception with no signs of RAM or data corruption issues. I test every new machine with Memtest86+. At some point I migrated the storage from Ext4 on LVMRAID to ZFS. All good. Then I wrote an alarm for Prometheus to tell me if there’s any issues in ZFS. A week later I get an email about a ZFS error. I check the system - says checksum errors, data has been corrected, applications unaffected, run a scrub to clear. I ran a scrub. A few more checksum errors found, all corrected, we’re clean now. There was a strong solar storm around that time, probably that. A couple of weeks later I get another email. Same symptoms, same procedure. No solar storm. Shit. Memtest86+ - pass. Hm. A couple of weeks later I get another. Same thing. Memtest again - nothing. This went on for several months. Meanwhile the off-site backup sees nothing like that. While running Memtest on another machine I noticed that the test passes following the first took longer than the first, a lot longer. I thought something might be wrong with that machine. Dug into it, got into Memtest’s source code and discovered that the first pass is shorter on purpose so that it quickly flags obviously bad RAM. Apparently if you want to detect less obvious issues, you have to run multiple passes. OK. Memtest the main server again, pass 1: OK, pass 2: OK, pass 3: OK, pass 4: FAIL. FUCK. Memtest each stick separately for 4 passes: OK. Memtest 2 at a time: OK. Memtest all 4: FAIL. Alright, now we know why ZFS keeps finding checksum errors. Long story short, this machine could not run this RAM in 4-DIMM config. Replaced it with another RAM that’s rated to run in 4-DIMM config on that processor. No more checksum issues. If I was running the older Ext4-on-LVMRAID storage stack, I would have caught NONE of these and it would have happily corrupted files here and there. In fact it likely did and I have some corruption. Moral of the story - run <em>many</em> Memtest passes and use <em>checksumming</em> storage stack like ZFS or Btrfs. I strongly recommend ZFS since its stripe RAID works fine unlike Btrfs’es. Start today, even if you’re working with a single disk and add redundancy when you can. Only after change Syncthing for something else if you still somehow get corruption without ZFS’es knowledge. And if ZFS tells you that you have checksum errors, you likely have bad hardware.</p>\n","inReplyTo":"https://lemmy.world/post/30402224","mediaType":"text/html","source":{"content":"That's really weird. I've been using it for mobile-desktop-server-offsite sync for many years, with transfer sizes over 15TB, over WiFi, cellular, cable, fiber. I've never seen data corruption. Conflicts, sometimes. Permission issues, sometimes. Wiping something accidentally, sometimes. It's even more weird because Syncthing performs computes hash values for the files it manages. I don't know if it performs hash validation after copying remotely but if not, it can be forced manually which would tell you what's fucked and be pulled from the source, if it still exists.\n\nNevermind, [it verifies the result:](https://docs.syncthing.net/users/syncing.html)\n\n> When a block is copied or received from another device, its SHA256 hash is computed and compared with the expected value. If it matches the block is written to a temporary copy of the file, otherwise it is discarded and Syncthing tries to find another source for the block.\n\nAccording to this, if you have data corruption it can only occur between copying/moving a temporary file on your destination to another directory, or it could occur on the source itself. Both of those scenarios are a cause of concern and would likely persist with any utility. Moving or copying a file from one location to another on a sane machine should not corrupt it. If I were you I'd ensure my server doesn't eat bits. If not the storage media, it could be bit rot, or bad RAM.\n\nJust in case everything seems fine, let me tell you what I dealt with. I had a Ryzen 5950X machine with 32GB of RAM. It worked well since inception with no signs of RAM or data corruption issues. I test every new machine with Memtest86+. At some point I migrated the storage from Ext4 on LVMRAID to ZFS. All good. Then I wrote an alarm for Prometheus to tell me if there's any issues in ZFS. A week later I get an email about a ZFS error. I check the system - says checksum errors, data has been corrected, applications unaffected, run a scrub to clear. I ran a scrub. A few more checksum errors found, all corrected, we're clean now. There was a strong solar storm around that time, probably that. A couple of weeks later I get another email. Same symptoms, same procedure. No solar storm. Shit. Memtest86+ - pass. Hm. A couple of weeks later I get another. Same thing. Memtest again - nothing. This went on for several months. Meanwhile the off-site backup sees nothing like that. While running Memtest on another machine I noticed that the test passes following the first took longer than the first, a lot longer. I thought something might be wrong with that machine. Dug into it, got into Memtest's source code and discovered that the first pass is shorter on purpose so that it quickly flags obviously bad RAM. Apparently if you want to detect less obvious issues, you have to run multiple passes. OK. Memtest the main server again, pass 1: OK, pass 2: OK, pass 3: OK, pass 4: FAIL. FUCK. Memtest each stick separately for 4 passes: OK. Memtest 2 at a time: OK. Memtest all 4: FAIL. Alright, now we know why ZFS keeps finding checksum errors. Long story short, this machine could not run this RAM in 4-DIMM config. Replaced it with another RAM that's rated to run in 4-DIMM config on that processor. No more checksum issues. If I was running the older Ext4-on-LVMRAID storage stack, I would have caught NONE of these and it would have happily corrupted files here and there. In fact it likely did and I have some corruption. Moral of the story - run *many* Memtest passes and use *checksumming* storage stack like ZFS or Btrfs. I strongly recommend ZFS since its stripe RAID works fine unlike Btrfs'es. Start today, even if you're working with a single disk and add redundancy when you can. Only after change Syncthing for something else if you still somehow get corruption without ZFS'es knowledge. And if ZFS tells you that you have checksum errors, you likely have bad hardware.","mediaType":"text/markdown"},"published":"2025-05-28T20:29:39.335302Z","updated":"2025-05-28T20:56:05.287078Z","tag":[{"href":"https://lemmy.world/u/ZeDoTelhado","name":"@ZeDoTelhado@lemmy.world","type":"Mention"}],"distinguished":false,"language":{"identifier":"en","name":"English"},"audience":"https://lemmy.world/c/selfhosted","attachment":[]},"cc":["https://lemmy.world/c/selfhosted","https://lemmy.world/u/ZeDoTelhado"],"tag":[{"href":"https://lemmy.world/u/ZeDoTelhado","name":"@ZeDoTelhado@lemmy.world","type":"Mention"}],"type":"Update","audience":"https://lemmy.world/c/selfhosted"},"cc":["https://lemmy.world/c/selfhosted/followers"],"type":"Announce","id":"https://lemmy.world/activities/announce/update/3e19197c-7f1d-4c5d-a58f-338cb39261a8"}
Response
Response Headers
Header | Value |
---|---|
cache-control | "no-cache, private" |
content-type | "application/activity+json" |
date | "Wed, 28 May 2025 20:56:18 GMT" |
x-debug-token | "578be4" |
Cookies
Request Cookies
No request cookies
Response Cookies
No response cookies
Session 1
Session Metadata
No session metadata
Session Attributes
No session attributes
Session Usage
1
Usages
Stateless check enabled
Usage |
---|
Symfony\Component\Security\Core\Authentication\Token\Storage\UsageTrackingTokenStorage:41
[ [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-core/Authentication/Token/Storage/UsageTrackingTokenStorage.php" "line" => 41 "function" => "getMetadataBag" "class" => "Symfony\Component\HttpFoundation\Session\Session" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-http/Authenticator/RememberMeAuthenticator.php" "line" => 69 "function" => "getToken" "class" => "Symfony\Component\Security\Core\Authentication\Token\Storage\UsageTrackingTokenStorage" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-http/Authentication/AuthenticatorManager.php" "line" => 111 "function" => "supports" "class" => "Symfony\Component\Security\Http\Authenticator\RememberMeAuthenticator" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-http/Firewall/AuthenticatorManagerListener.php" "line" => 34 "function" => "supports" "class" => "Symfony\Component\Security\Http\Authentication\AuthenticatorManager" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-http/Authenticator/Debug/TraceableAuthenticatorManagerListener.php" "line" => 40 "function" => "supports" "class" => "Symfony\Component\Security\Http\Firewall\AuthenticatorManagerListener" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-bundle/Debug/WrappedLazyListener.php" "line" => 38 "function" => "supports" "class" => "Symfony\Component\Security\Http\Authenticator\Debug\TraceableAuthenticatorManagerListener" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-http/Firewall/AbstractListener.php" "line" => 25 "function" => "supports" "class" => "Symfony\Bundle\SecurityBundle\Debug\WrappedLazyListener" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-bundle/Security/LazyFirewallContext.php" "line" => 60 "function" => "__invoke" "class" => "Symfony\Component\Security\Http\Firewall\AbstractListener" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-bundle/Debug/TraceableFirewallListener.php" "line" => 80 "function" => "__invoke" "class" => "Symfony\Bundle\SecurityBundle\Security\LazyFirewallContext" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/security-http/Firewall.php" "line" => 95 "function" => "callListeners" "class" => "Symfony\Bundle\SecurityBundle\Debug\TraceableFirewallListener" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/event-dispatcher/Debug/WrappedListener.php" "line" => 116 "function" => "onKernelRequest" "class" => "Symfony\Component\Security\Http\Firewall" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/event-dispatcher/EventDispatcher.php" "line" => 220 "function" => "__invoke" "class" => "Symfony\Component\EventDispatcher\Debug\WrappedListener" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/event-dispatcher/EventDispatcher.php" "line" => 56 "function" => "callListeners" "class" => "Symfony\Component\EventDispatcher\EventDispatcher" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/event-dispatcher/Debug/TraceableEventDispatcher.php" "line" => 139 "function" => "dispatch" "class" => "Symfony\Component\EventDispatcher\EventDispatcher" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/http-kernel/HttpKernel.php" "line" => 157 "function" => "dispatch" "class" => "Symfony\Component\EventDispatcher\Debug\TraceableEventDispatcher" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/http-kernel/HttpKernel.php" "line" => 76 "function" => "handleRaw" "class" => "Symfony\Component\HttpKernel\HttpKernel" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/http-kernel/Kernel.php" "line" => 197 "function" => "handle" "class" => "Symfony\Component\HttpKernel\HttpKernel" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/symfony/runtime/Runner/Symfony/HttpKernelRunner.php" "line" => 35 "function" => "handle" "class" => "Symfony\Component\HttpKernel\Kernel" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/vendor/autoload_runtime.php" "line" => 29 "function" => "run" "class" => "Symfony\Component\Runtime\Runner\Symfony\HttpKernelRunner" "type" => "->" ] [ "file" => "/var/www/kbin/kbin/public/index.php" "line" => 7 "args" => [ "/var/www/kbin/kbin/vendor/autoload_runtime.php" ] "function" => "require_once" ] ] |
Flashes
Flashes
No flash messages were created.
Server Parameters
Server Parameters
Defined in .env
Key | Value |
---|---|
APP_ENV | "dev" |
APP_SECRET | "82ce1339a6c267e28d1f1dcb37a7454c" |
CORS_ALLOW_ORIGIN | "^https?://(kbin.localhost|127\.0\.0\.1)(:[0-9]+)?$" |
DATABASE_URL | "postgresql://kbin:917eaa3d703f19d123@127.0.0.1:5433/kbin?serverVersion=15&charset=utf8" |
HCAPTCHA_SECRET | "" |
HCAPTCHA_SITE_KEY | "" |
JWT_PASSPHRASE | "" |
JWT_PUBLIC_KEY | "%kernel.project_dir%/config/jwt/public.pem" |
JWT_SECRET_KEY | "%kernel.project_dir%/config/jwt/private.pem" |
KBIN_ADMIN_ONLY_OAUTH_CLIENTS | "false" |
KBIN_API_ITEMS_PER_PAGE | "25" |
KBIN_CAPTCHA_ENABLED | "false" |
KBIN_CONTACT_EMAIL | "kbin@j0h.nl" |
KBIN_DEFAULT_LANG | "en" |
KBIN_DOMAIN | "kbin.spritesserver.nl" |
KBIN_FEDERATION_ENABLED | "true" |
KBIN_FEDERATION_PAGE_ENABLED | "true" |
KBIN_HEADER_LOGO | "false" |
KBIN_JS_ENABLED | "true" |
KBIN_META_DESCRIPTION | "a private kbin install" |
KBIN_META_KEYWORDS | "kbin, content agregator, open source, fediverse" |
KBIN_META_TITLE | "Sprites kbin instance" |
KBIN_REGISTRATIONS_ENABLED | "true" |
KBIN_SENDER_EMAIL | "kbin@j0h.nl" |
KBIN_STORAGE_URL | "https://kbin.spritesserver.nl/media/" |
KBIN_TITLE | "/kbin" |
LOCK_DSN | "flock" |
MAILER_DSN | "smtp://spritesmods.com" |
MERCURE_JWT_SECRET | "231e9a1277f5585d52aa0b1e34c0f984xxxx" |
MERCURE_PUBLIC_URL | "https://kbin.spritesserver.nl/.well-known/mercure" |
MERCURE_URL | "http://localhost:3000/.well-known/mercure" |
MESSENGER_TRANSPORT_DSN | "doctrine://default" |
OAUTH_FACEBOOK_ID | "" |
OAUTH_FACEBOOK_SECRET | "" |
OAUTH_GITHUB_ID | "" |
OAUTH_GITHUB_SECRET | "" |
OAUTH_GOOGLE_ID | "" |
OAUTH_GOOGLE_SECRET | "" |
POSTGRES_DB | "kbin" |
POSTGRES_PASSWORD | "917eaa3d703f19d123" |
POSTGRES_USER | "kbin" |
POSTGRES_VERSION | "15" |
REDIS_DNS | "redis://uSJBDOQfuOMgt8kyGhpUzViTnQSEdEJTsOIsYSsg3v40v@localhost" |
REDIS_PASSWORD | "uSJBDOQfuOMgt8kyGhpUzViTnQSEdEJTsOIsYSsg3v40v" |
S3_BUCKET | "media.karab.in" |
S3_KEY | "" |
S3_REGION | "eu-central-1" |
S3_SECRET | "" |
S3_VERSION | "latest" |
Defined as regular env variables
Key | Value |
---|---|
APP_DEBUG | "1" |
CONTENT_LENGTH | "9531" |
CONTENT_TYPE | "application/activity+json" |
CONTEXT_DOCUMENT_ROOT | "/var/www/kbin/kbin/public" |
CONTEXT_PREFIX | "" |
DOCUMENT_ROOT | "/var/www/kbin/kbin/public" |
GATEWAY_INTERFACE | "CGI/1.1" |
HTTPS | "on" |
HTTP_ACCEPT | "*/*" |
HTTP_ACCEPT_ENCODING | "gzip" |
HTTP_DATE | "Wed, 28 May 2025 20:56:17 GMT" |
HTTP_DIGEST | "SHA-256=cSiBieXOkr0ygkVhKE9Nr9xgXhY+r3yoW8O6fuH9ZV4=" |
HTTP_HOST | "kbin.spritesserver.nl" |
HTTP_SIGNATURE | "keyId="https://lemmy.world/c/selfhosted#main-key",algorithm="hs2019",headers="(request-target) content-type date digest host",signature="FklvhoDl/5vnM0fzslD+hvd93MwpjjKGDDVZwsKiVuTuaKOk7eRqmAHaQKbY2oeocNvd3WTEuVhgNZBDs6tK5u2oMShl0caAmqgxh7yL90oAP73U2jVbArO7nOpk4/NRIpK41abPhnMFUIT/73SnpvoIGLhhPZXgwXtBn3ZdWa6DJt/4FHTvvZ2CDNfbilANFPn29yMPfJJ7Av7bhJcYzLVw2xc9yD5rcqtRwTpNcREcgnHkiT5Ut63cOaS8zxzmKDRZTwGKf3dkPTbi3Lzo56xkFfW1oXUUXMhjIAYSC3NzLClJyS1cHqBmioGIAWXtllR5dzcGl9csc6fi+/LS3g=="" |
HTTP_TRACEPARENT | "00-e89f33937cf52c0e58d9770c68da3405-b599710561ead89a-01" |
HTTP_TRACESTATE | "" |
HTTP_USER_AGENT | "Lemmy/0.19.11-19-g2895f45e8; +https://lemmy.world" |
PATH | "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin" |
PHP_SELF | "/index.php" |
QUERY_STRING | "" |
REDIRECT_HTTPS | "on" |
REDIRECT_SSL_CIPHER | "TLS_AES_256_GCM_SHA384" |
REDIRECT_SSL_CIPHER_ALGKEYSIZE | "256" |
REDIRECT_SSL_CIPHER_EXPORT | "false" |
REDIRECT_SSL_CIPHER_USEKEYSIZE | "256" |
REDIRECT_SSL_CLIENT_VERIFY | "NONE" |
REDIRECT_SSL_COMPRESS_METHOD | "NULL" |
REDIRECT_SSL_PROTOCOL | "TLSv1.3" |
REDIRECT_SSL_SECURE_RENEG | "true" |
REDIRECT_SSL_SERVER_A_KEY | "rsaEncryption" |
REDIRECT_SSL_SERVER_A_SIG | "sha256WithRSAEncryption" |
REDIRECT_SSL_SERVER_I_DN | "CN=R10,O=Let's Encrypt,C=US" |
REDIRECT_SSL_SERVER_I_DN_C | "US" |
REDIRECT_SSL_SERVER_I_DN_CN | "R10" |
REDIRECT_SSL_SERVER_I_DN_O | "Let's Encrypt" |
REDIRECT_SSL_SERVER_M_SERIAL | "06E2DDD4A973DA92D92F46873A8653408A8E" |
REDIRECT_SSL_SERVER_M_VERSION | "3" |
REDIRECT_SSL_SERVER_SAN_DNS_0 | "kbin.spritesserver.nl" |
REDIRECT_SSL_SERVER_S_DN | "CN=kbin.spritesserver.nl" |
REDIRECT_SSL_SERVER_S_DN_CN | "kbin.spritesserver.nl" |
REDIRECT_SSL_SERVER_V_END | "Jun 28 21:01:47 2025 GMT" |
REDIRECT_SSL_SERVER_V_START | "Mar 30 21:01:48 2025 GMT" |
REDIRECT_SSL_SESSION_ID | "5af76a4991c2f56e12411f92606be9471987af93d1241e583c1c4bb695bbd69a" |
REDIRECT_SSL_SESSION_RESUMED | "Initial" |
REDIRECT_SSL_TLS_SNI | "kbin.spritesserver.nl" |
REDIRECT_SSL_VERSION_INTERFACE | "mod_ssl/2.4.62" |
REDIRECT_SSL_VERSION_LIBRARY | "OpenSSL/3.0.16" |
REDIRECT_STATUS | "200" |
REDIRECT_URL | "/f/inbox" |
REMOTE_ADDR | "135.181.143.221" |
REMOTE_PORT | "57848" |
REQUEST_METHOD | "POST" |
REQUEST_SCHEME | "https" |
REQUEST_TIME | 1748465777 |
REQUEST_TIME_FLOAT | 1748465777.6137 |
REQUEST_URI | "/f/inbox" |
SCRIPT_FILENAME | "/var/www/kbin/kbin/public/index.php" |
SCRIPT_NAME | "/index.php" |
SERVER_ADDR | "5.9.62.165" |
SERVER_ADMIN | "webmaster@spritesmods.com" |
SERVER_NAME | "kbin.spritesserver.nl" |
SERVER_PORT | "443" |
SERVER_PROTOCOL | "HTTP/1.1" |
SERVER_SIGNATURE | "" |
SERVER_SOFTWARE | "Apache" |
SSL_CIPHER | "TLS_AES_256_GCM_SHA384" |
SSL_CIPHER_ALGKEYSIZE | "256" |
SSL_CIPHER_EXPORT | "false" |
SSL_CIPHER_USEKEYSIZE | "256" |
SSL_CLIENT_VERIFY | "NONE" |
SSL_COMPRESS_METHOD | "NULL" |
SSL_PROTOCOL | "TLSv1.3" |
SSL_SECURE_RENEG | "true" |
SSL_SERVER_A_KEY | "rsaEncryption" |
SSL_SERVER_A_SIG | "sha256WithRSAEncryption" |
SSL_SERVER_I_DN | "CN=R10,O=Let's Encrypt,C=US" |
SSL_SERVER_I_DN_C | "US" |
SSL_SERVER_I_DN_CN | "R10" |
SSL_SERVER_I_DN_O | "Let's Encrypt" |
SSL_SERVER_M_SERIAL | "06E2DDD4A973DA92D92F46873A8653408A8E" |
SSL_SERVER_M_VERSION | "3" |
SSL_SERVER_SAN_DNS_0 | "kbin.spritesserver.nl" |
SSL_SERVER_S_DN | "CN=kbin.spritesserver.nl" |
SSL_SERVER_S_DN_CN | "kbin.spritesserver.nl" |
SSL_SERVER_V_END | "Jun 28 21:01:47 2025 GMT" |
SSL_SERVER_V_START | "Mar 30 21:01:48 2025 GMT" |
SSL_SESSION_ID | "5af76a4991c2f56e12411f92606be9471987af93d1241e583c1c4bb695bbd69a" |
SSL_SESSION_RESUMED | "Initial" |
SSL_TLS_SNI | "kbin.spritesserver.nl" |
SSL_VERSION_INTERFACE | "mod_ssl/2.4.62" |
SSL_VERSION_LIBRARY | "OpenSSL/3.0.16" |
SYMFONY_DOTENV_VARS | "KBIN_DOMAIN,KBIN_TITLE,KBIN_DEFAULT_LANG,KBIN_FEDERATION_ENABLED,KBIN_CONTACT_EMAIL,KBIN_SENDER_EMAIL,KBIN_JS_ENABLED,KBIN_REGISTRATIONS_ENABLED,KBIN_API_ITEMS_PER_PAGE,KBIN_STORAGE_URL,KBIN_META_TITLE,KBIN_META_DESCRIPTION,KBIN_META_KEYWORDS,KBIN_HEADER_LOGO,KBIN_CAPTCHA_ENABLED,KBIN_FEDERATION_PAGE_ENABLED,REDIS_PASSWORD,REDIS_DNS,S3_KEY,S3_SECRET,S3_BUCKET,S3_REGION,S3_VERSION,OAUTH_FACEBOOK_ID,OAUTH_FACEBOOK_SECRET,OAUTH_GOOGLE_ID,OAUTH_GOOGLE_SECRET,OAUTH_GITHUB_ID,OAUTH_GITHUB_SECRET,KBIN_ADMIN_ONLY_OAUTH_CLIENTS,APP_ENV,APP_SECRET,POSTGRES_DB,POSTGRES_USER,POSTGRES_PASSWORD,POSTGRES_VERSION,DATABASE_URL,MESSENGER_TRANSPORT_DSN,MAILER_DSN,MERCURE_URL,MERCURE_PUBLIC_URL,MERCURE_JWT_SECRET,CORS_ALLOW_ORIGIN,LOCK_DSN,JWT_SECRET_KEY,JWT_PUBLIC_KEY,JWT_PASSPHRASE,HCAPTCHA_SITE_KEY,HCAPTCHA_SECRET" |