Ubuntu - No space left on device Is it a lie or have I run out of inodes

Yesterday one of my development servers decided it was going to do some very strange things.Wordpress and other websites stopped working properly, I got session errors when trying to usePHPMyAdmin, I couldn't upload files through web forms (the server complained there was no temporary directory). So I logged in to try and work out what was going on. The temporary directory was there and had the correct permissions, however if I tried to create a file in it I was told:

$ touch /tmp/testfile

Unable to create file /tmp/testfile: No space left on device

So I must have run out of disk space, which is odd as I had loads last time I checked.

$ df -h

Filesystem            Size  Used Avail Use% Mounted on

/dev/sda1              15G  8.5G  6.5G  57% /

devtmpfs              299M  112K  299M   1% /dev

none                  308M     0  308M   0% /dev/shm

none                  308M   64K  308M   1% /var/run

none                  308M     0  308M   0% /var/lock

none                  308M     0  308M   0% /lib/init/rw

/dev/sdc1              40G  6.4G   32G  17% /home

Oh, I have plenty of disk space! What the hell is going on then? As my server is an Amazon EC2instance my first thoughts were there was a problem with the block storage. So I spent an hour or so trying to find any clues in their forums and got nowhere.

After another few hours of scouring the internet for people having similar problems and finding nothing at all I was about to give up. As a last ditch attempt to find the solution I checked myMunin stats for the server and immediately I noticed that the inode graphs for one of the mounted disks had been rising steadily over the last few weeks and had just reached 100%!!!

$ df -i

Filesystem            Inodes   IUsed   IFree IUse% Mounted on

/dev/sda1             983040  983040       0  100% /

devtmpfs               76490    1957   74533    3% /dev

none                   78747       1   78746    1% /dev/shm

none                   78747      34   78713    1% /var/run

none                   78747       2   78745    1% /var/lock

none                   78747       1   78746    1% /lib/init/rw

/dev/sdc1            2621440   13238 2608202    1% /home

So then, where are all these files? There must be thousands of them to be using up 100% of just under a million.

To count all the files in a directory and all it's subdirectories:

$ for i in /*; do echo $i; find $i | wc -l; done

Then you can narrow down your search by replacing the /* for any directory that has an unusually large number of files in. For me it was /var

$ for i in /var/*; do echo $i; find $i | wc -l; done

Eventually I narrowed it down to the reports being held by the Squid Proxy server report generator sarg so a simple fix was to clear out all the old reports and stop sarg from auto generating reports every day.

$ rm -rf /var/log/sarg/*

And thats it! Server fixed and back up and running without any problems. All I have to do is remember to keep an eye on any autogenerated logs and reports and make sure that old ones are actually being deleted!

 

posted @ 2013-05-21 10:47  zaleilynn  阅读(656)  评论(0编辑  收藏  举报