Posts

Showing posts from December, 2023

AutoMine® Concept - the next generation of mining automation - Behind th...

Image

GPT Pilot ✈️ Build Full Stack Apps with a SINGLE PROMPT (Made for Devs)

Image

NEOVIM CONF 2023 - Introducing nvim-llama

Image

Artificial Intelligence, Automation, Work, and Algorithms | Matteo Pasqu...

Image

"Ruthless" Pricing, AI APIs Pricing gets a CRAZY drop!!!

Image

The BIG Mistral AI Secret is OUT :) (And I'm very happy)!!!

Image

Mamba: Linear-Time Sequence Modeling with Selective State Spaces

 Foundation models, now powering most of the exciting applications in deep learning, are almost universally based on the Transformer architecture and its core attention module. Many subquadratic-time architectures such as linear attention, gated convolution and recurrent models, and structured state space models (SSMs) have been developed to address Transformers’ computational inefficiency on long sequences, but they have not performed as well as attention on important modalities such as language. We identify that a key weakness of such models is their inability to perform content-based reasoning, and make several improvements. First, simply letting the SSM parameters be functions of the input addresses their weakness with discrete modalities, allowing the model to selectively propagate or forget information along the sequence length dimension depending on the current token. Second, even though this change prevents the use of efficient convolutions, we design a hardware-aware paral...

Interview with Dr. Ilya Sutskever, co-founder of OPEN AI - at the Open U...

Image

How Transformers ended an age old tradition in Machine Learning (Explain...

Image

GEMINI Beats GPT4!! Google's New Gemini Model Is INSANE 🤯

Image

BEST 2 out of 7 LLMs for Causal Reasoning (7B-70B)

Image

CUDA Developer Tools | SOL Analysis with NVIDIA Nsight Compute

Image
Image
 

Deep Progressive Techno #19

Image