Skip to main content

Claude 2, Google's NotebookLM, LongNet, Quivr

Anthropic's new Claude model, LongNet & scaling to a billion tokens, and extracting information from unstructured data with Quivr!
Created on July 12|Last edited on July 12

Claude 2

Anthropic just released Claude 2 and can be tested here. It not only boasts substantially better performance than the original Claude model, but it also supports a max prompt length of 100k tokens. Another unique feature, from one of their demo videos, is their app's ability to upload files! Including files in the conversation is as seamless as dragging and dropping.

Check it out today!

NotebookLM

Google's new NotebookLM utilizes LLMs as your personal note-taking assistant. It can be used for summarization, Q&A, and generating ideas. Join the waitlist now!

LongNet

The paper "LongNet: Scaling Transformers to 1,000,000,000 Tokens" introduces dilated attention which, simply put, divides the Q, K, V matrices of shape (N,d)(N, d) in vanilla attention into ww equal length matrices. For each of these (Qi,Ki,Vi)(Q_i, K_i, V_i), an interval rr is selected to sparsely select certain rows of these matrices. The authors describe dilated attention as a drop-in replacement for standard attention, trading off the globality of the original attention mechanism for computational efficiency without sacrificing performance on shorter sequences. Unlike attention, dilated attention doesn't grow quadratically to the input sequence length, allowing for a huge scale up to, as the title puts it, a billion tokens.


Quivr

Quivr, a recently trending GitHub repository, leverages LLMs to extract information from unstructured data. They support a variety of modalities:
  • Text
  • Markdown
  • PDF
  • Powerpoint
  • Excel (Not Yet)
  • CSV
  • Word
  • Audio
  • Video
Go check out their demo and app!

References

Tags: ML News
Iterate on AI agents and models faster. Try Weights & Biases today.