Jupyter notebook for using LlamaIndex with arXiv papers for retrieval-augmented generation (RAG).
This paper presents a detailed vocabulary of 33 terms and a taxonomy of 58 LLM prompting techniques, along with guidelines for prompt engineering and a meta-analysis of natural language prefix-prompting, serving as the most comprehensive survey on prompt engineering to date.
The article discusses the competition Nvidia faces from Intel and AMD in the GPU market. While these competitors have introduced new accelerators that match or surpass Nvidia's offerings in terms of memory capacity, performance, and price, Nvidia maintains a strong advantage through its CUDA software ecosystem. CUDA has been a significant barrier for developers switching to alternative hardware due to the effort required to port and optimize existing code. However, both Intel and AMD have developed tools to ease this transition, like AMD's HIPIFY and Intel's SYCL. Despite these efforts, the article notes that the majority of developers now write higher-level code using frameworks like PyTorch, which can run on different hardware with varying levels of support and performance. This shift towards higher-level programming languages has reduced the impact of Nvidia's CUDA moat, though challenges still exist in ensuring compatibility and performance across different hardware platforms.
Sparse autoencoders (SAEs) have been trained on Llama 3.3 70B, releasing an interpreted model accessible via API, enabling research and product development through feature space exploration and steering.
A set of open-source models and tools that enables any developer to build a state-of-the-art AI application with full control over data privacy, cost, and performance.
Improve your online learning experience with the Britannica AI Chatbot, powered by a comprehensive database of encyclopedic articles. Ask questions and receive reliable answers across various topics.
A workshop hosted by Princeton Language and Intelligence featuring experts on developing, evaluating, and ensuring the reliability of AI agents, with sessions on developer tools, real-world evaluations, and ensuring reliability.
MemryX has launched a $149 M.2 module designed to provide AI processing capabilities for compact systems, offering up to 24 TOPS of compute power and supporting various data formats.
This article discusses methods to measure and improve the accuracy of Large Language Model (LLM) applications, focusing on building an SQL Agent where precision is crucial. It covers setting up the environment, creating a prototype, evaluating accuracy, and using techniques like self-reflection and retrieval-augmented generation (RAG) to enhance performance.
MarkItDown is a utility for converting various files to Markdown, including PDF, PowerPoint, Word, Excel, Images, Audio, HTML, text-based formats, and ZIP files.