LLMII uses a local LLM to label metadata and index images. It does not rely on a cloud service or database. A visual language model runs on your computer and is used to create captions and keywords for images in a directory tree. The generated information is then added to each image file's metadata.