LLM 0.17 release enables multi-modal input, allowing users to send images, audio, and video files to Large Language Models like GPT-4o, Llama, and Gemini, with a Python API and cost-effective pricing.
Simon Willison recently delivered a talk during the Mastering LLMs: A Conference For Developers & Data Scientists, which was a six-week long online event. The talk centered around Simon's LLM Python command-line utility and its plugins, emphasizing how they can be utilized to explore Large Language Models (LLMs) and perform various tasks. Last week, he discussed accessing LLMs from the command-line, sharing valuable insights and techniques with the audience.
The api_base key can be used to point the OpenAI client library at a different API endpoint.
The author has also automated their weeknotes by using an Observable notebook, which generates the "releases this week" and "TILs this week" sections.
The notebook fetches TILs from the author's Datasette, grabs releases from GitHub, and assembles a markdown string for the new post.
* `llm` CLI tool for running prompts against large language models
* Automation of weeknotes using an Observable notebook
* Notebook generates "releases this week" and "TILs this week" sections
* Tool stores prompts and responses in a SQLite database