I know that the ‘insufficient disk space’ issue has been raised before but I still have problems.
I’m trying to crawl a file share with >500,000 documents, initially I hit an issue with /boithoData filling up so I resized the disk to 100Gb.
The crawl now runs to a point (around 8500 documents) before it stops with the ‘insufficient disk space’ message, although df-h still shows 99% free space in /boithoData. After a short time the crawl will start again, it takes a long time to get back to the point it was at before crawling another 2000 or so documents before giving the ‘insufficient disk space’ message again.
The crawl has now been running 24hrs and is only up to 430,000 documents and /boithoData is showing 94% free - is this how long a crawl such as this should take? Would each subsequent crawl take just as long or is it just the initial crawl that takes time? I’d really like to get to the bottom of this as Searchdaimon looks like a great product!
Any suggestions?