#transformer-models

[ follow ]
fromHackernoon
8 months ago

Even AI Needs Glasses: When Space Images Get Too Fuzzy to Fix | HackerNoon

Although our Transformer-based deep learning model provides state-of-the-art performance in both resolution enhancement and noise reduction for moderate noise levels, restoration becomes impossible when the noise level exceeds a threshold.
Data science
fromThegreenplace
1 month ago

Sparsely-gated Mixture Of Experts (MoE)

The feed forward layer in transformer models is crucial for reasoning on token relationships, often housing most of the model's weights due to its larger dimensionality.
Marketing tech
#machine-learning
Artificial intelligence
fromHackernoon
1 year ago

New AI Method Lets Models Decide What to Think About | HackerNoon

Mixture-of-Depths Transformers improve efficiency in transformer architectures by dynamically allocating computational resources.
Artificial intelligence
fromHackernoon
1 year ago

New AI Method Lets Models Decide What to Think About | HackerNoon

Mixture-of-Depths Transformers improve efficiency in transformer architectures by dynamically allocating computational resources.
fromHackernoon
1 year ago

Evaluating the Performance of vLLM: How Did It Do? | HackerNoon

The evaluation of vLLM was conducted using models with various parameters, specifically targeting configurations that reflect popular sizes in the LLM landscape like those of GPT-3.
Data science
fromHackernoon
9 months ago

Where does In-context Translation Happen in Large Language Models: Inference Efficiency | HackerNoon

The potential of speeding up transformer inference lies in identifying where task recognition occurs in the model, which helps in optimizing processing and reducing redundancy.
Data science
[ Load more ]