On daily basis, I only work with a static site which has 1K pages. But then I need to export more than 100.000 markdown files to a static website.

As before, I choose Hugo to build the static site, which I think it should be fast and light enough to process that. But.. hm, not exactly as I thought.. 🤡

So, here are some taken points.

hugo serve

The first problem that I got was no response when I run hugo serve, it was because building the 100K+ markdown files was not easy and light process.

So, forget about preview 100K pages, get some samples of few files or a few hundred files, only to test and preview the site.


My first response of not responded hugo serve command actually stop the process (ctrl c) then run hugo command to build the markdown files. Of course, this command has produced the same result, not responded.

To debug, try to add debug parameter hugo --debug and then it started to show the building process one by one for thousands of template rendering.

After waiting for about 3 hours with a almost overhead notebook, finally, the build process was done successfully 😀

Upload to Firebase hosting, and it takes almost 2GB storage, including the static images and documents. First deployment success ☑️.


I start to evaluate on the Google Webmaster Tools and I saw that the sitemap got rejected because it was more than 50K or URLs. Here’s 1 the reference for maximum 50MB and 50.000 URLs.

Try to find the solution in hugo discussion board 23, but currently no easy solution to split it. May be because not so many sites will produce more than 50K URLs.

I did not want to spend to much time with this problem, so I just use this 4 python script to split the sitemap xml file.

delete 100K+ files in single directory

Sometimes I need to clean up the markdown files, but the linux rm command unable to delete a list of more than 100.000 file names, because the argument was too long.

So I need to use a pipe to xargs rm command:

find contents/ -name "*.md" | xargs rm

set lower resource usages to rebuild the static site

As I mentioned before, the build process took about 3 hours with 8 CPUs usages, which triggered a high temperature in my notebook. Next, I am planning to build on small VPS instances, or even Github actions. So, I need to simulate the process to set lower resource usages.

My first try was to set a lower priority of the hugo process by running command nice hugo, but I saw that it’s not effective, even though the resource usages getting lower.

2nd try was using docker to limit the CPU usage, and it worked very well. Here’s the command that I used:

HUGO_ENV=production docker run --rm -it --cpus="1.5" -v $(pwd):/src klakegg/hugo:0.74.3 --cleanDestinationDir --minify --debug

I used minimal docker images from 5 and docker runtime options from [6].

I took about 7 hours to complete the hugo build process with 1.5 CPUs. Sorry, can not share the metrics.

That’s all folks!

Photo by Clark Gu on Unsplash.

  1. https://support.google.com/webmasters/answer/183668?hl=en&ref_topic=4581190 ↩︎

  2. https://discourse.gohugo.io/t/please-add-limit-the-sitemap-size/20421 ↩︎

  3. https://discourse.gohugo.io/t/limitation-for-sitemap-xml-size/2726 ↩︎

  4. https://gist.github.com/kravietz/f2eec690ecb4dde84206 ↩︎

  5. https://github.com/klakegg/docker-hugo ↩︎