Skip to content

trestad/CogAttn

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ee705b5 · Nov 13, 2024

History

5 Commits
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 13, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024
Nov 11, 2024

Repository files navigation

Official PyTorch implementation of "More Expressive Attention with Negative Weights"

TL;DR: We propose a novel attention mechanism, named Cog Attention, that enables attention weights to be negative for enhanced expressiveness.

Why named cog attention?

  1. The attention pattern looks like cogs.
  2. The transformation cog ("T-cog") and the living metal of each transformer's body allows them to change from their natural robotic body into an "alternate mode" based on something a form of technology or life form that they've observed and scanned. —— Wikipedia. In summary, the cog enhances the expressiveness of Transformers :) net_path=YOUR_PATH_SAVE_MY_CKPT/cog_nnet_ema.pth &> cfg_mscoco_sample_cog.log &

We have to withdraw the open-source code temporarily as we are currently undergoing a code review in the company, and we will re-open source it soon.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages