This n8n template shows you how to automate document summarization while keeping full digital sovereignty. By combining Nextcloud for file storage and the IONOS AI Model Hub, your sensitive documents stay on European infrastructure—fully outside US CLOUD Act jurisdiction.
A Schedule Trigger kicks off the workflow at regular intervals. It calls the Nextcloud List a Folder node to retrieve all files in your chosen directory.
A Code node then acts as a filter, checking each file's last-modified timestamp and letting through only files changed within the past 24 hours.
Filtered files are pulled into the workflow via the Nextcloud Download a File node. From there, the Summarization Chain takes over: a Default Data Loader and Token Splitter prepare the content, and the IONOS AI Model Hub Chat Model generates a concise, structured summary.
The final output is saved back to your Nextcloud instance as a note for your team.
Nextcloud setup: Make sure your Nextcloud credentials include both read and write permissions—one to download source files, one to upload the generated summary.
Model selection: The IONOS AI Model Hub currently offers several LLMs, including Llama 3.1 8B, Mistral Nemo 12B, Mistral Small 24B, Llama 3.3 70B, GPT-OSS 120B, and Llama 3.1 405B. For document summarization, Llama 3.3 70B strikes the best balance between output quality and speed. If you're processing high volumes or need faster turnaround, Mistral Nemo 12B is a leaner alternative.
24 value to change how far back the workflow looks.