ChronoStriker1

Members
  • Posts

    148
  • Joined

  • Last visited

Converted

  • Gender
    Undisclosed

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

ChronoStriker1's Achievements

Apprentice

Apprentice (3/14)

5

Reputation

  1. It was the cpu and Intel has RMA'ed it. Unfortunately the usb drive that I was using seems to have failed so I now have a licensing support ticket open with unraid, hopefully that can be taken care of relatively quickly.
  2. Nope still unstable with the same kind of errors, guess contacting intel is next
  3. OK ive replaced the board, so far it "feels" more stable but Im still seeing segfaults (it has not hard crashed on me as of yet). Anyone mind talking a look at the diagnostics and a quick grab of my current syslog?
  4. The motherboard seems like its going to be the hardest thing to rma since I cant find how to do an Advanced RMA through ASUS and as much as Im crashing I dont want to be without the server for at least 10 days. Related though since I dont think its the ram i decided to just buy the ram that was sent as a replacement so I have 128GB now, which is nice.
  5. Atomic burn worked without issue max temp was 71C and no crashes while it did it.
  6. CPU temps have been fine this entire time. Ill install the plugin and run it after the latest parity check is complete. I will also contact intel and asus to see if I can rma the processor and motherboard sicne im outside my 30 day window with Amazon.
  7. Well that will be fun to rma then. Another question that is hopefully easy, even after reboots and swapping the ram, its always the same things that always segfault. Wouldn't random things segfault as it thinks there is an issue or it runs out of memory? Looking at my syslog from yesterday: May 9 21:37:29 Tower kernel: unraid-api[16221]: segfault at ffffffffffff3b28 ip 0000000001518f00 sp 00007ffe4d6f11a8 error 5 in unraid-api[91c000+167b000] likely on CPU 0 (core 0, socket 0) May 9 21:45:25 Tower kernel: python3[11316]: segfault at 7 ip 00001504488506f3 sp 00007ffdc92cd5f0 error 4 in libpython3.10.so.1.0[15044873b000+1be000] likely on CPU 14 (core 28, socket 0) May 9 22:25:13 Tower kernel: Thunar[22347]: segfault at 600000003a ip 00001512f15d1f1c sp 00007ffc678c45b0 error 4 in libglib-2.0.so.0.6600.8[1512f157e000+88000] likely on CPU 12 (core 24, socket 0) May 9 22:25:18 Tower kernel: thunar[1956]: segfault at 600000003a ip 0000153cad537f1c sp 00007ffc9fc1f040 error 4 in libglib-2.0.so.0.6600.8[153cad4e4000+88000] likely on CPU 2 (core 4, socket 0) May 9 22:45:28 Tower kernel: python[24692]: segfault at 1 ip 00001507b28ac411 sp 00007fff2a585a50 error 6 in libpython3.11.so.1.0[1507b2799000+1bb000] likely on CPU 0 (core 0, socket 0) May 9 23:56:34 Tower kernel: unraid-api[8794]: segfault at ffffffffffff3b28 ip 0000000001518f00 sp 00007fff36d19508 error 5 in unraid-api[91c000+167b000] likely on CPU 0 (core 0, socket 0) May 10 01:09:03 Tower kernel: unraid-api[15814]: segfault at ffffffffffff3b28 ip 0000000001518f00 sp 00007fffca0cda58 error 5 in unraid-api[91c000+167b000] likely on CPU 0 (core 0, socket 0) May 10 02:37:20 Tower kernel: python[7637]: segfault at 8 ip 000014f6acd47ac9 sp 000014f6a84bba90 error 4 in libpython3.9.so.1.0[14f6acc13000+1b8000] likely on CPU 0 (core 0, socket 0) May 10 03:14:47 Tower kernel: python3[27555]: segfault at 0 ip 000014dc983af61b sp 000014dc95621998 error 6 in libpython3.8.so.1.0[14dc98273000+183000] likely on CPU 12 (core 24, socket 0) May 10 03:46:44 Tower kernel: python[4413]: segfault at 6 ip 0000151abbf715e6 sp 00007ffc2b350e40 error 6 in libpython3.11.so.1.0[151abbe5f000+1bb000] likely on CPU 0 (core 0, socket 0) May 10 05:24:00 Tower kernel: php7[4270]: segfault at 40 ip 00005585e53dd3a0 sp 00007ffc2b656380 error 4 in php7[5585e5200000+240000] likely on CPU 0 (core 0, socket 0) May 10 05:29:39 Tower kernel: unraid-api[30617]: segfault at ffffffffffff3b28 ip 0000000001518f00 sp 00007ffe8bc8fca8 error 5 in unraid-api[91c000+167b000] likely on CPU 0 (core 0, socket 0) May 10 06:32:13 Tower kernel: unraid-api[21391]: segfault at ffffffffffff3b28 ip 0000000001518f00 sp 00007ffe80db5a58 error 5 in unraid-api[91c000+167b000] likely on CPU 0 (core 0, socket 0) I know for a fact that unraid-api, python3, and thunar are always programs (or the libriaries associated with them) that seem to segfault. Is it possible that some of the files have been damaged due to the crashes and thats why they are faulting?
  8. Welp tried one stick twice had the same issue, was able to get another set and am having the same errors, so I think at this point I can say its not the ram. So where would the next place to check be?
  9. And it looks like it crashed again. I can attempt running one stick at a time later today to see if there is any changes but is there anything else I can test other than just the memory?
  10. I have disabled xmp, there has been at least one segfault that I noticed so far but it hasnt crashed yet. I will continue to keep an eye on it.
  11. After another crash yesterday I ran a full 4 pass memtest86 run and my memory passed. Well it's still possible the issue is the memory I need to be more specific so I can RMA parts. Is there any better way to track down what's going on.
  12. The system stopped letting me do some actions again, one cpu looked like it was pegged at 100% by "/usr/src/app/vendor/bundle/ruby/3.1.0/bin/rake jobs:work" tried killing it but it wouldnt die, I tried stoping things in order to reboot but the web interface became unresponsive, I attempted to reboot from the commandline but the last message I saw was "Tower init: Trying to re-exec init". It did eventually reboot but it had an uncleen shutdown.
  13. I had run it last night for 1 pass and the ram had passed. I can attempt to let it run over night sometime this weekend.
  14. After deleting the file and doing another scrub that error went away. I also manually moved thigns again, this time arround it looks like things moved (it looked like it stopped part way through previously). Currently I am not getting that shfs: cache disk full message. Still waiting for it to crash again.