Solaris is one of the most stable operating systems available… but lets face it, stuff happens. Solaris does panic, but I want everyone to be clear, a “panic”, despite the seemingly contradictory name, is by its nature a controlled event. When the kernel encounters behavior that is uncorrectable and will cause irreparable harm to the running system or, even worse, corrupt data, the system will voluntarily tap out using the panic system call to get the system down quickly, hopefully leaving a core dump in its wake for post-mortem analysis.
In this blog entry we’ll discuss core dumps and panic’s in general. In part 2 we’ll discuss a tool to make life just a little easier, the Solaris Crash Analysis Tool, or “Solaris CAT”.
I want to point out that post-mortem core analysis is really the task of a kernel engineer. The fact is, way less than 1% of us who ever engage in core analysis are actually going to have any real idea of what the hell we’re doing. And thats ok! I guarantee that you’ll post something from an analysis to a mailing list and you’ll get some asshole who forgets that he’s been paid to work on the Solaris kernel for the last 20 years while you work a job which is now on hold because of said core dump, with replies like “We can clearly see that due to the memory address in this register that you are a moron….” The point here is, if you don’t know what your doing, don’t be discouraged. What we, mere mortals, are trying to do is not necessarily solve the problem but provide clues which will help us guide our search, either by posting a stack trace to a mailing list, or send the dump to Sun Support, or to take a panic string and search the bug database or Google for. The cuddletech rule of crashes is:
An unexpected crash is unacceptable; An unexplained crash is inexcusable.
If you’re reading this you’ve probably lived through a panic before, but lets recap. The best explanation of a “crash” event and resulting dump can be found in the dumpadm(1M) man page:
A crash dump is a disk copy of the physical memory of the computer at the time of a fatal system error. When a fatal operating system error occurs, a message describing the error is printed to the console. The operating system then generates a crash dump by writing the contents of phy- sical memory to a predetermined dump device, which is typi- cally a local disk partition. The dump device can be config- ured by way of dumpadm. Once the crash dump has been writ- ten to the dump device, the system will reboot. Fatal operating system errors can be caused by bugs in the operating system, its associated device drivers and loadable modules, or by faulty hardware. Whatever the cause, the crash dump itself provides invaluable information to your support engineer to aid in diagnosing the problem. As such, it is vital that the crash dump be retrieved and given to your support provider. Following an operating system crash, the savecore(1M) utility is executed automatically during boot to retrieve the crash dump from the dump device, and write it to a pair of files in your file system named unix.X and vmcore.X, where X is an integer identifying the dump. Together, these data files form the saved crash dump. The directory in which the crash dump is saved on reboot can also be configured using dumpadm.
I encourage you to read both the savecore(1M) and dumpadm(1M) man pages. You’ll find that with savecore -L you can create a dump of a live system, so if you don’t have a crashed system around to play with, use that. Alternatively, you can use reboot -d to dump a core and reboot.
At this point we’ll assume you have a dump available. By default you’ll find them in /var/crash/hostname/, you’ll have dumps in pairs: vmcore.0 and unix.0. We feed these two files to mdb, the (-k, kernel) Modular DeBugger, to preform our analysis like so:
# mdb -k unix.0 vmcore.0 Loading modules: [ unix krtld genunix specfs dtrace cpu.AuthenticAMD.15 uppc pcplusmp ufs ip sctp usba lofs zfs random ipc md fcip fctl fcp crypto logindmux ptm nfs ] >
You are now free to move about the dump. mdb commands are strange and unusual at first, it takes a lot of time to get comfortable with it, but there are a couple of debugger commands that can give us the essence of what we need. Lets walk through them.
The ::status command will display high level information regarding this debugging session. Of usefulness here is the dumps “panic message” and OS release.
> ::status debugging crash dump vmcore.0 (64-bit) from hostname operating system: 5.11 snv_43 (i86pc) panic message: BAD TRAP: type=e (#pf Page fault) rp=fffffe80000ad3d0 addr=0 occurred in module "unix" due to a NULL pointer dereference dump content: kernel pages only
The ::stack command will prove you with a stack trace, this is the same thing trace you would have seen in syslog or the console.
> ::stack atomic_add_32() nfs_async_inactive+0x55(fffffe820d128b80, 0, ffffffffeff0ebcb) nfs3_inactive+0x38b(fffffe820d128b80, 0) fop_inactive+0x93(fffffe820d128b80, 0) vn_rele+0x66(fffffe820d128b80) snf_smap_desbfree+0x78(fffffe8185e2ff60) dblk_lastfree_desb+0x25(fffffe817a30f8c0, ffffffffac1d7cc0) dblk_decref+0x6b(fffffe817a30f8c0, ffffffffac1d7cc0) freeb+0x89(fffffe817a30f8c0) tcp_rput_data+0x215f(ffffffffb4af7140, fffffe812085d780, ffffffff993c3c00) squeue_enter_chain+0x129(ffffffff993c3c00, fffffe812085d780, fffffe812085d780, 1, 1) ip_input+0x810(ffffffffa23eec68, ffffffffaeab8040, fffffe812085d780, e) i_dls_link_ether_rx_promisc+0x266(ffffffff9a4c35f8, ffffffffaeab8040, fffffe812085d780) mac_rx+0x7a(ffffffffa2345c40, ffffffffaeab8040, fffffe812085d780) e1000g_intr+0xf6(ffffffff9a4b2000) av_dispatch_autovect+0x83(1a) intr_thread+0x50()
The ::msgbuf command will output the message buffer at the time of crash; the message buffer is most commonly used by sysadmins through the “dmesg” command.
> ::msgbuf MESSAGE .... WARNING: IP: Hardware address '00:14:4f:xxxxxxx' trying to be our address xxxx WARNING: IP: Hardware address '00:14:4f:xxxx' trying to be our address xxxx panic[cpu0]/thread=fffffe80000adc80: BAD TRAP: type=e (#pf Page fault) rp=fffffe80000ad3d0 addr=0 occurred in module "unix" due to a NULL pointer dereference sched: #pf Page fault Bad kernel fault at addr=0x0 .... blah blah, snipped for brevity.
The ::panicinfo command will give you lots of fun cryptic counter information, of most interest is the first 3 lines, which contain the CPU on which the panic occured, the running thread, and the panic message. You’ll notice these are commonly repeated and the most useful pieces of information.
> ::panicinfo cpu 0 thread fffffe80000adc80 message BAD TRAP: type=e (#pf Page fault) rp=fffffe80000ad3d0 addr=0 occurred in module "unix" due to a NULL pointer dereference rdi 0 rsi 1 rdx fffffe80000adc80 rcx 0 r8 0 r9 fffffe80dba125c0 rax 0 rbx fffffe8153a36040 rbp fffffe80000ad4e0 r10 3e0 r10 3e0 r11 ffffffffaeab8040 r12 ffffffffb7b4cac0 r13 0 r14 fffffe820d128b80 r15 ffffffffeff0ebcb fsbase ffffffff80000000 gsbase fffffffffbc27850 ds 43 es 43 fs 0 gs 1c3 trapno e err 2 rip fffffffffb838680 cs 28 rflags 10246 rsp fffffe80000ad4c8 ss 0 gdt_hi 0 gdt_lo defacedd idt_hi 0 idt_lo 80300fff ldt 0 task 60 cr0 80050033 cr2 0 cr3 10821b000
In my opinion, the koolest command is ::cpuinfo -v. Truth be told, if you run multiple applications on a server the most common question people (especially managers) want answered is “which application did it?”, being translated into geek-esse “who do I blame?” This command will help you determine that by displaying, complete with beautiful ASCII art, the threads and process names running on each CPU (NRUN). In the following example, we know the event occured on CPU 0, thus thats the one we want to look at. Note that the “sched” process should be interpreted as “kernel”.
> ::cpuinfo -v ID ADDR FLG NRUN BSPL PRI RNRN KRNRN SWITCH THREAD PROC 0 fffffffffbc2f370 1b 1 0 165 no no t-1 fffffe80000adc80 sched | | | RUNNING <--+ | +--> PIL THREAD READY | 6 fffffe80000adc80 EXISTS | - fffffe80daab6a20 ruby ENABLE | +--> PRI THREAD PROC 99 fffffe8000b88c80 sched ID ADDR FLG NRUN BSPL PRI RNRN KRNRN SWITCH THREAD PROC 1 ffffffff983b3800 1f 1 0 59 yes no t-0 fffffe80daac2f20 smtpd | | RUNNING <--+ +--> PRI THREAD PROC READY 99 fffffe8000bacc80 sched QUIESCED EXISTS ENABLE ID ADDR FLG NRUN BSPL PRI RNRN KRNRN SWITCH THREAD PROC 2 ffffffff9967a800 1f 2 0 -1 no no t-0 fffffe8000443c80 (idle) | | RUNNING <--+ +--> PRI THREAD PROC READY 99 fffffe8000b82c80 sched QUIESCED 60 fffffe80018f8c80 sched EXISTS ENABLE ID ADDR FLG NRUN BSPL PRI RNRN KRNRN SWITCH THREAD PROC 3 ffffffff9967a000 1f 1 0 -1 no no t-0 fffffe8000535c80 (idle) | | RUNNING <--+ +--> PRI THREAD PROC READY 60 fffffe8000335c80 zsched QUIESCED EXISTS ENABLE
The ::ps command allows us to see all running processes. Several flags are supported, including -z to display Zone ID’s.
> ::ps -z S PID PPID PGID SID ZONE UID FLAGS ADDR NAME R 0 0 0 0 0 0 0x00000001 fffffffffbc25900 sched R 3 0 0 0 0 0 0x00020001 ffffffff9970d928 fsflush R 2 0 0 0 0 0 0x00020001 ffffffff9970e558 pageout R 1 0 0 0 0 0 0x42004000 ffffffff9970f188 init R 20534 1 20533 20533 24 1006 0x42010400 ffffffffb246f9b8 ruby R 20532 1 20531 20531 24 1006 0x42010400 fffffe8109674308 ruby R 20529 1 20528 20528 24 1006 0x42010400 fffffe80dc5602f0 ruby ...
We can use ::pgrep to search for processes and use the appropriate address for further digging. In the following example I’ll find a Java process and then determine which zone that process was running in:
> ::pgrep java S PID PPID PGID SID UID FLAGS ADDR NAME R 3628 1 3620 3574 0 0x42004400 fffffe80deeb3240 java > fffffe80deeb3240::print proc_t p_zone->zone_name p_zone->zone_name = 0xffffffffae0cef00 "testzone03"
There are many more tools and way to dig into your dumps using mdb. It can be confusing because you need to reference things by address, but you get more comfortable with it as you play around. If you are interested in learning more I highly recommend reading Eric Lowe’s “Examining the Anatomy of a Process”, which digs into the topic of process examination via mdb.
One thing you’ll notice in all this is that the messages at the time of crash on the console or in syslog contain almost everything you need to know without digging too deeply. Therefore, assuming you have those messages, the most useful thing most people will extract from the core files is the output of the ::cpuinfo command to see what process was on the offending CPU at the time of the crash. Knowing what processes, zones, etc, were running at the time of crash are interesting but rarely mean much if they weren’t directly involved in the panic.
As I said, once you start getting into referencing memory addresses to deepen your analysis things get sticky and tricky very very quickly… thats where Solaris CAT comes in, which we’ll talk about in part 2.