All Projects → d-li14 → lambda.pytorch

d-li14 / lambda.pytorch

Licence: other
PyTorch implementation of Lambda Network and pretrained Lambda-ResNet

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to lambda.pytorch

keras cv attention models
Keras/Tensorflow attention models including beit,botnet,CMT,CoaT,CoAtNet,convnext,cotnet,davit,efficientdet,efficientnet,fbnet,gmlp,halonet,lcnet,levit,mlp-mixer,mobilevit,nfnets,regnet,resmlp,resnest,resnext,resnetd,swin,tinynet,uniformer,volo,wavemlp,yolor,yolox
Stars: ✭ 159 (+194.44%)
Mutual labels:  imagenet, attention
Adversarial-Examples-in-PyTorch
Pytorch code to generate adversarial examples on mnist and ImageNet data.
Stars: ✭ 112 (+107.41%)
Mutual labels:  imagenet
PyTorch-Model-Compare
Compare neural networks by their feature similarity
Stars: ✭ 119 (+120.37%)
Mutual labels:  imagenet
stagin
STAGIN: Spatio-Temporal Attention Graph Isomorphism Network
Stars: ✭ 34 (-37.04%)
Mutual labels:  attention
gnn-lspe
Source code for GNN-LSPE (Graph Neural Networks with Learnable Structural and Positional Representations), ICLR 2022
Stars: ✭ 165 (+205.56%)
Mutual labels:  attention
image-classification
A collection of SOTA Image Classification Models in PyTorch
Stars: ✭ 70 (+29.63%)
Mutual labels:  imagenet
gqa-node-properties
Recalling node properties from a knowledge graph
Stars: ✭ 19 (-64.81%)
Mutual labels:  attention
protein-transformer
Predicting protein structure through sequence modeling
Stars: ✭ 77 (+42.59%)
Mutual labels:  attention
alexnet-architecture.tensorflow
Unofficial TensorFlow implementation of "AlexNet" architecture.
Stars: ✭ 15 (-72.22%)
Mutual labels:  imagenet
LNSwipeCell
一套友好的、方便集成的针对cell的左滑编辑功能!
Stars: ✭ 16 (-70.37%)
Mutual labels:  attention
TF-NAS
TF-NAS: Rethinking Three Search Freedoms of Latency-Constrained Differentiable Neural Architecture Search (ECCV2020)
Stars: ✭ 66 (+22.22%)
Mutual labels:  imagenet
OpenPrompt
An Open-Source Framework for Prompt-Learning.
Stars: ✭ 1,769 (+3175.93%)
Mutual labels:  pre-trained-model
LFattNet
Attention-based View Selection Networks for Light-field Disparity Estimation
Stars: ✭ 41 (-24.07%)
Mutual labels:  attention
ShapleyExplanationNetworks
Implementation of the paper "Shapley Explanation Networks"
Stars: ✭ 62 (+14.81%)
Mutual labels:  iclr2021
OASIS
Official implementation of the paper "You Only Need Adversarial Supervision for Semantic Image Synthesis" (ICLR 2021)
Stars: ✭ 232 (+329.63%)
Mutual labels:  iclr2021
datastories-semeval2017-task6
Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (-62.96%)
Mutual labels:  attention
super-gradients
Easily train or fine-tune SOTA computer vision models with one open source training library
Stars: ✭ 429 (+694.44%)
Mutual labels:  imagenet
image-recognition
采用深度学习方法进行刀具识别。
Stars: ✭ 19 (-64.81%)
Mutual labels:  attention
Relation-Extraction-Transformer
NLP: Relation extraction with position-aware self-attention transformer
Stars: ✭ 63 (+16.67%)
Mutual labels:  attention
NIPS-Global-Paper-Implementation-Challenge
Selective Classification For Deep Neural Networks.
Stars: ✭ 11 (-79.63%)
Mutual labels:  imagenet

lambda.pytorch

[NEW!] Check out our latest work involution in CVPR'21 that bridges convolution and self-attention operators.


PyTorch implementation of LambdaNetworks: Modeling long-range Interactions without Attention.

Lambda Networks apply associative law of matrix multiplication to reverse the computing order of self-attention, achieving the linear computation complexity regarding content interactions.

Similar techniques have been used previously in A2-Net and CGNL. Check out a collection of self-attention modules in another repository dot-product-attention.

Training Configuration

✓ SGD optimizer, initial learning rate 0.1, momentum 0.9, weight decay 0.0001

✓ epoch 130, batch size 256, 8x Tesla V100 GPUs, LR decay strategy cosine

✓ label smoothing 0.1

Pre-trained checkpoints

Architecture Parameters FLOPs Top-1 / Top-5 Acc. (%) Download
Lambda-ResNet-50 14.995M 6.576G 78.208 / 93.820 model | log

Citation

If you find this repository useful in your research, please cite

@InProceedings{Li_2021_CVPR,
author = {Li, Duo and Hu, Jie and Wang, Changhu and Li, Xiangtai and She, Qi and Zhu, Lei and Zhang, Tong and Chen, Qifeng},
title = {Involution: Inverting the Inherence of Convolution for Visual Recognition},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2021}
}
@inproceedings{
bello2021lambdanetworks,
title={LambdaNetworks: Modeling long-range Interactions without Attention},
author={Irwan Bello},
booktitle={International Conference on Learning Representations},
year={2021},
}
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].