Back to News
Advertisement
Advertisement

⚑ Community Insights

Discussion Sentiment

50% Positive

Analyzed from 153 words in the discussion.

Trending Topics

#more#data#context#llm#right#gives#model#mathematically#shape#reminds

Discussion (3 Comments)Read Original on HackerNews

vorticalboxβ€’about 4 hours ago
This reminds me of https://dnhkng.github.io/posts/rys/

David looks into the LLM finds the thinking layers and cut duplicates then and put them back to back.

This increases the LLM scores with basically no over head.

Very interesting read.

renticulousβ€’about 1 hour ago
Jeff Dean says models hallucinate because their training data is "squishy."

But what's in the context window is sharp, the exact text or video frame right in front of them.

The goal is to bring more of the world into that context.

Compression gives it intuition. Context gives it precision.

Imagine if we could extract the model's reasoning core and plug it anywhere we want.

kangβ€’40 minutes ago
The answer should be obvious that its both.

Zurada was one of our AI textbook that makes it visual that right from a simple classifier to a large language model, we are mathematically creating a shape(, that the signal interacts with). More parameters would mean shape can be curved in more ways and more data means the curve is getting hi-definition.

They reach something with data, treating neural network as blackbox, which could be derived mathematically using the information we know.