Ask the llm to summarize stuff when it's running low, should be better than just rememberig the first bit and a window.
AI Jason
100K subscribersIt's hard to get LLM generate big amount of content and take in large inputs; To solve this, introducing StreamingLLM, Extend ...
I don't understand how can it help even for books. Will...
22 Comments
Wes Roth
4.7k views • 15 hours ago
TheAIGRID
11.2k views • 1 day ago
TheAIGRID
31.4k views • 1 day ago
Wes Roth
39.6k views • 1 day ago
TheAIGRID
22.7k views • 3 days ago
Wes Roth
51.0k views • 4 days ago
TheAIGRID
26.5k views • 4 days ago
TheAIGRID
9.2k views • 5 days ago
TheAIGRID
35.6k views • 6 days ago
Wes Roth
62.8k views • 6 days ago
Shelf will be hidden for 30 daysUndo
TheAIGRID
45.1k views • 6 days ago
Wes Roth
27.1k views • 1 week ago
Wes Roth
55.4k views • 1 week ago
AI Jason
81.4k views • 1 week ago
TheAIGRID
36.6k views • 1 week ago
TheAIGRID
6.3k views • 1 week ago
TheAIGRID
19.6k views • 1 week ago
TheAIGRID
13.6k views • 1 week ago
Wes Roth
51.9k views • 1 week ago
AI For Humans
598 views • 1 week ago
Wes Roth
60.9k views • 1 week ago
TheAIGRID
147.6k views • 1 week ago
TheAIGRID
42.4k views • 1 week ago
AI For Humans
3.0k views • 2 weeks ago
TheAIGRID
30.8k views • 2 weeks ago
Wes Roth
48.9k views • 2 weeks ago
Wes Roth
39.1k views • 2 weeks ago
TheAIGRID
32.0k views • 2 weeks ago
Wes Roth
59.6k views • 2 weeks ago
TheAIGRID
33.5k views • 2 weeks ago
TheAIGRID
85.7k views • 2 weeks ago
Wes Roth
92.3k views • 2 weeks ago
Morningside AI
8.0k views • 2 weeks ago
Wes Roth
139.6k views • 2 weeks ago
AI For Humans
763 views • 2 weeks ago
TheAIGRID
28.6k views • 2 weeks ago
Wes Roth
33.4k views • 2 weeks ago
AI Explained
123.1k views • 2 weeks ago
TheAIGRID
457.7k views • 3 weeks ago
AI For Humans
1.4k views • 3 weeks ago
TheAIGRID
21.7k views • 3 weeks ago
Wes Roth
28.5k views • 3 weeks ago
Wes Roth
48.6k views • 3 weeks ago
TheAIGRID
398.9k views • 3 weeks ago
Wes Roth
19.7k views • 3 weeks ago
TheAIGRID
31.3k views • 3 weeks ago
AI Jason
38.3k views • 3 weeks ago
Wes Roth
60.8k views • 3 weeks ago
Wes Roth
20.6k views • 3 weeks ago
TheAIGRID
18.1k views • 3 weeks ago
TheAIGRID
53.2k views • 3 weeks ago
Wes Roth
46.5k views • 3 weeks ago
Wes Roth
33.5k views • 3 weeks ago
Wes Roth
83.3k views • 3 weeks ago
Wes Roth
35.4k views • 3 weeks ago
TheAIGRID
40.1k views • 3 weeks ago
Wes Roth
37.4k views • 3 weeks ago
Wes Roth
16.2k views • 3 weeks ago
TheAIGRID
26.4k views • 3 weeks ago
AI Explained
115.9k views • 3 weeks ago
AI For Humans
2.9k views • 4 weeks ago
TheAIGRID
41.8k views • 4 weeks ago
TheAIGRID
40.3k views • 4 weeks ago
AI Jason
68.2k views • 1 month ago
TheAIGRID
40.6k views • 1 month ago
AI For Humans
373 views • 1 month ago
TheAIGRID
51.3k views • 1 month ago
TheAIGRID
19.7k views • 1 month ago
TheAIGRID
31.6k views • 1 month ago
TheAIGRID
103.6k views • 1 month ago
AI For Humans
3.5k views • 1 month ago
TheAIGRID
45.8k views • 1 month ago
TheAIGRID
56.0k views • 1 month ago
AI Explained
115.3k views • 1 month ago
AI For Humans
2.3k views • 1 month ago
AI For Humans
1.7k views • 1 month ago
AI Jason
27.4k views • 1 month ago
AI For Humans
334 views • 1 month ago
AI For Humans
2.6k views • 1 month ago
AI Explained
105.5k views • 1 month ago
AI Explained
130.2k views • 1 month ago
AI For Humans
1.5k views • 1 month ago
AI Jason
203.2k views • 1 month ago
AI For Humans
1.4k views • 1 month ago
AI Explained
177.3k views • 2 months ago
AI Jason
33.0k views • 2 months ago
AI Explained
150.6k views • 2 months ago
Andrej Karpathy
453.8k views • 2 months ago
AI Explained
240.7k views • 2 months ago
AI Jason
62.2k views • 2 months ago
AI Explained
187.2k views • 2 months ago
AI Explained
160.9k views • 3 months ago
AI Jason
87.8k views • 3 months ago
AI Explained
270.8k views • 3 months ago
AI Jason
60.7k views • 3 months ago
AI Explained
96.5k views • 3 months ago
AI Jason
7.1k views • 3 months ago
AI Explained
145.7k views • 3 months ago
AI Explained
133.0k views • 4 months ago
AI Explained
79.4k views • 4 months ago
AI Jason
16.2k views • 4 months ago
AI Explained
84.0k views • 4 months ago
AI Explained
74.5k views • 4 months ago
AI Explained
144.7k views • 5 months ago
AI Jason
68.7k views • 5 months ago
Morningside AI
4.1k views • 5 months ago
AI Explained
83.7k views • 5 months ago
AI Jason
134.4k views • 5 months ago
AI Jason
33.3k views • 5 months ago
Morningside AI
9.7k views • 5 months ago
AI Explained
229.0k views • 5 months ago
Andrej Karpathy
1.8M views • 5 months ago
AI Explained
112.8k views • 5 months ago
AI Explained
167.2k views • 5 months ago
AI Explained
156.8k views • 5 months ago
Morningside AI
26.0k views • 6 months ago
AI Jason
16.1k views • 6 months ago
AI Explained
96.4k views • 6 months ago
AI Jason
68.8k views • 6 months ago
AI Explained
120.6k views • 6 months ago
AI Jason
52.6k views • 6 months ago
AI Jason
19.1k views • 7 months ago
AI Jason
52.7k views • 7 months ago
AI Jason
28.6k views • 7 months ago
AI Jason
13.6k views • 7 months ago
AI Jason
183.6k views • 8 months ago
AI Jason
48.3k views • 8 months ago
AI Jason
26.5k views • 8 months ago
AI Jason
14.4k views • 8 months ago
Andrej Karpathy
4.2M views • 1 year ago
Andrej Karpathy
153.6k views • 1 year ago
Andrej Karpathy
167.4k views • 1 year ago
Andrej Karpathy
240.3k views • 1 year ago
Andrej Karpathy
269.9k views • 1 year ago
22 Comments
Ask the llm to summarize stuff when it's running low, should be better than just rememberig the first bit and a window.
🎯 Key Takeaways for quick navigation:
00:00 📚 Increasing Data Input for Large Language Models
- Large languag     See More
wait wait! you just finished too quickly I hope you have elaborated more on what can't be accomplished I didn't quite get why and the difference between what we can achieve with long     See More
Which LLM has the largest token limit to expand the context length of the chat?
This is my third video I’ve seen today from you and you are so consistent with providing value with your words. Thank you my new AI guru🙏
I don't understand how can it help even for books. Will it forget everything in the middle of the book? I try to think how it works in human brain. When we reed a book we (usually) don     See More
Great job at providing information about new developments, Jason! Thanks!
Ask the llm to summarize stuff when it's running low, should be better than just rememberig the first bit and a window.     See Less
🎯 Key Takeaways for quick navigation:
00:00 📚 Increasing Data Input for Large Language Models
- Large languag     See More ace challenges with increased data input.
- GPU memory limitations and computational time impact performance.
- The concept of "Window attention" has been used to mitigate these issues.
01:09 🔄 Introducing StreamingLM and Attention Sync
- StreamingLM is a research project to enhance data input for large language models.
- Attention Sync focuses on the importance of initial tokens for context.
- StreamingLM combines initial tokens and a rolling cache for effective context.
02:49 🔓 Unlocking Possibilities with StreamingLM
- StreamingLM enables handling long-form content generation and movie transcripts.
- It works well for scenarios that require generating a large amount of content.
- However, it may not handle extremely complex tasks with extensive context loss.
Made with HARPA AI    See Less
is there is still no solution for extending data?     See Less
wait wait! you just finished too quickly I hope you have elaborated more on what can't be accomplished I didn't quite get why and the difference between what we can achieve with long     See More ents.    See Less
as usual, inspiring, accurate and updated. ty sir     See Less
Which LLM has the largest token limit to expand the context length of the chat?     See Less
This is my third video I’ve seen today from you and you are so consistent with providing value with your words. Thank you my new AI guru🙏     See Less
i luv u     See Less
I don't understand how can it help even for books. Will it forget everything in the middle of the book? I try to think how it works in human brain. When we reed a book we (usually) don     See More ber each word. What we do, we create visual images inside and it compress the book very effectively. Supposedly, these images are like tokens or maybe like embeddings and don't occupy much space in memory. Is it possible to implement something like this for LLMs? They should kind of learn during "reading the book" and they should convert texts to multimodal embeddings or even find (create) approximate path in embedding space and later they should have the ability to analyze this path later. Not sure how it should be implemented.    See Less
Great job at providing information about new developments, Jason! Thanks!     See Less