vikunja, plane, taiga, focalboard, wekan, restyaboard, planka
integrated in: openproject, anytype, cryptpad, nextcloud, php version of owncloud
vikunja, plane, taiga, focalboard, wekan, restyaboard, planka
integrated in: openproject, anytype, cryptpad, nextcloud, php version of owncloud
Hmm i can confirm that this is not working. I wonder if the new support only is for one "primary" OIDC provider. There is also nothing popping up when using matrix.org :( – but when using their testserver, it just redirects and doesn't even show a button.
I found this discussion which might cover your issue: https://community.vikunja.io/t/repeatable-tasks-appear-to-just-have-start-due-dates-reset-instead-of-creating-a-clone/
Have you changed the homeserver to tchncs.de
? I don't know (but would assume) if it can autodetect this... . Oh by the way, if you used matrix.tchncs.de
directly, the information for sliding sync was not configured there yet, sorry. I have added this now.
Hmm if people really care about this, sure. Will try to deploy within this week (but no promises).
I am a bit confused now... the spare was 98% as to read in my snippet above ... where does it say "no spare available"? I think it is on me to request a swap, and thats what i did as also the one with slightly less wear reported 255% used – which afaik is an aprox. lifetime left estimation based on rw cycles (not sure about all factors).
The one the hoster left in for me to play with, said no:
[Wed Jul 26 19:19:10 2023] nvme nvme1: I/O 9 QID 0 timeout, disable controller
[Wed Jul 26 19:19:10 2023] nvme nvme1: Device shutdown incomplete; abort shutdown
[Wed Jul 26 19:19:10 2023] nvme nvme1: Removing after probe failure status: -4
Tried multiple kernelflags n stuff but couldn't get past that error. Would have been interesting to have the hoster ship the thing to me (and maybe that would have been a long enough cooldown to have the thing working again), but i assume that would have been expensive from helsinki.
because i wanted to give it some exposure for a few hours :)
i am not sure about those numbers on the new ones ... it was one db restore and a few hrs of uptime ... a scrub... , then i rsynced some stuff over and since then the thing is in idle 🤷
sample of the current active system .. i think at time of arrival it was 2+tb written or something
SMART/Health Information (NVMe Log 0x02)
Critical Warning: 0x00
Temperature: 37 Celsius
Available Spare: 100%
Available Spare Threshold: 10%
Percentage Used: 0%
Data Units Read: 88,116,921 [45.1 TB]
Data Units Written: 43,968,235 [22.5 TB]
Host Read Commands: 689,015,212
Host Write Commands: 409,762,513
Controller Busy Time: 1,477
Power Cycles: 4
Power On Hours: 248
Unsafe Shutdowns: 0
Media and Data Integrity Errors: 0
Error Information Log Entries: 0
Warning Comp. Temperature Time: 0
Critical Comp. Temperature Time: 0
Temperature Sensor 1: 37 Celsius
Temperature Sensor 2: 46 Celsius
Error Information (NVMe Log 0x01, 16 of 64 entries)
No Errors Logged
Yes, it is federated – however since there is no SSO on the Lemmy instance, you need to make a new account. Like you need to make new accounts between email providers. :) However it is a different federation protocol: Matrix vs ActivityPub. For more cool stuff, check out https://tchncs.de :3
Dang the old host was deleted from the monitoring – however looking on at least one smart thing from my emails, there were no errors logged before the drives gave up on life during replacement. They just had a ton read/written and the used counter at 255% (even tho rw and age were not equal, its weird and one reason why i wanted to have at least one replaced in the first place). This is the one that had more:
SMART/Health Information (NVMe Log 0x02)
Critical Warning: 0x04
Temperature: 53 Celsius
Available Spare: 98%
Available Spare Threshold: 10%
Percentage Used: 255%
Data Units Read: 7,636,639,249 [3.90 PB]
Data Units Written: 2,980,551,083 [1.52 PB]
Host Read Commands: 87,676,174,127
Host Write Commands: 28,741,297,023
Controller Busy Time: 705,842
Power Cycles: 7
Power On Hours: 17,437
Unsafe Shutdowns: 1
Media and Data Integrity Errors: 0
Error Information Log Entries: 0
Warning Comp. Temperature Time: 0
Critical Comp. Temperature Time: 0
Temperature Sensor 1: 53 Celsius
Temperature Sensor 2: 64 Celsius
Error Information (NVMe Log 0x01, 16 of 64 entries)
No Errors Logged
SMART/Health Information (NVMe Log 0x02)
Critical Warning: 0x00
Temperature: 24 Celsius
Available Spare: 100%
Available Spare Threshold: 5%
Percentage Used: 3%
Data Units Read: 122,135,021 [62.5 TB]
Data Units Written: 31,620,076 [16.1 TB]
Host Read Commands: 1,014,224,069
Host Write Commands: 231,627,064
Controller Busy Time: 3,909
Power Cycles: 2
Power On Hours: 117
Unsafe Shutdowns: 0
Media and Data Integrity Errors: 0
Error Information Log Entries: 4
Warning Comp. Temperature Time: 0
Critical Comp. Temperature Time: 0
Temperature Sensor 1: 24 Celsius
Error Information (NVMe Log 0x01, 16 of 256 entries)
Num ErrCount SQId CmdId Status PELoc LBA NSID VS
0 4 0 0x0000 0x8004 0x000 0 0 -
SMART/Health Information (NVMe Log 0x02)
Critical Warning: 0x00
Temperature: 24 Celsius
Available Spare: 100%
Available Spare Threshold: 5%
Percentage Used: 2%
Data Units Read: 153,193,333 [78.4 TB]
Data Units Written: 29,787,075 [15.2 TB]
Host Read Commands: 1,262,977,843
Host Write Commands: 230,135,280
Controller Busy Time: 4,804
Power Cycles: 11
Power On Hours: 119
Unsafe Shutdowns: 5
Media and Data Integrity Errors: 0
Error Information Log Entries: 14
Warning Comp. Temperature Time: 0
Critical Comp. Temperature Time: 0
Temperature Sensor 1: 24 Celsius
Error Information (NVMe Log 0x01, 16 of 256 entries)
Num ErrCount SQId CmdId Status PELoc LBA NSID VS
0 14 0 0x100d 0x8004 0x000 0 0 -
i understand that you are disappointed and it does not have to be a permanent decision. right now i lack time and energy to give it a close look again. earlier the instance did not look like something we should federate with based on its general content and our experience.
remember that its
user@org
– in this case@tchncs.de