Feature request: AI prompt management - poor tagging/labelling/rating results out of the box

I like the AI features that come with DEVONthink4, but I am somewhat underwhelmed by the results of AI tagging, AI labeling, and AI rating (although the rating is somewhat secondary).
One thing that I note in the application preferences is that only the summarization allows for a custom prompt; the other capabilities do not seem to be configurable.
Do you have a similar experience with ways to improve things without relying on the most expensive LLM?

I’m not sure what you’re expecting but any kind of automated metadata like this is non-configurable. If you have specific ideas about that should and shouldn’t be, you should handle these things on your own manually or work to perfect prompts that deliver results closer to what you hope to see.

1 Like

Which model did you actually use? This can make a huge difference.

Thanks for your thoughts, actually what you mention “…or work to perfect prompts that deliver results closer to what you hope to see.” is what I’m asking, I don’t see a way to modify the prompts that generate tagging, labeling, and rating.

I’ve tried ollama with small models, and indeed qwen3:0.6b is (as expected) giving very poor results. I moved to ChatGPT 4.1 nano and the results were better but still, in my view subpar. But it is also true that I don’t know what exactly is fed into the LLM, so the prompt might be of limited impact.

Tagging depends highly on personal preferences which might also change over time. E.g. nobody in this company would tag like I do and I don’t tag like I did in the past.

One possibility to improve automatic tagging is to create the desired tags first and then to enable the option to use only the existing tags, see Settings > Files > Tags

Another possibility is to use batch processing or smart rules, here’s a basic example:

1 Like

Thanks a lot, I will try this approach! Your help is highly appreciated!