Pytorch: Target Detection Network-FPN
Pytorch: Target Detection-Feature Pyramid-FPN
Copyright: Jingmin Wei, Pattern Recognition and Intelligent System, School of Artificial and Intelligence, Huazhong University of Science and Technology
Pytorch Tutorial Column Link
This tutorial is not commercial and is only for learning and reference exchange. If you need to reproduce it, ...
Added by magicmoose on Wed, 16 Feb 2022 19:24:48 +0200
FedAvg, the basic algorithm of Federated learning based on PyTorch
1. Foreword
In a previous blog Code implementation of federal learning basic algorithm FedAvg Using numpy hand built neural network to realize FedAvg, the effect of hand built neural network has been very excellent, not
II. Data introduction
There are multiple clients in federated learning. Each client has its own data set, which they are u ...
Added by nakins on Tue, 15 Feb 2022 17:59:41 +0200
Python identification verification code -- Digital beauty icon Click
Write in front
Recently, there was an uproar in the script circle when the news that a crowd coding platform was run away came out (I'm not in the script circle, I just like to watch the people in the group blow and force), as if I could never hear the familiar advertising language again. This also indicates that the third-party coding pla ...
Added by punk3d on Sun, 13 Feb 2022 16:51:03 +0200
Interpretation of longform code structure
While longformer extends maxlen, there are many structural difficulties, which are gradually analyzed here.
_ sliding_ chunks_ query_ key_ Structural transformation in matmul function
The hardest thing to understand here is these sentences
query_size = list(query.size())
query_size[1] = query_size[1]*2-1
query_stride = list(query.stride( ...
Added by throx on Sat, 12 Feb 2022 16:54:24 +0200
[machine learning] WGAN theory and code analysis based on pytoch
catalogue
1. There is a problem with the original GAN
2 WGAN principle
3 code understanding
GitHub source code
Reference article: Amazing Wasserstein GAN - Zhihu (zhihu.com)
1. There is a problem with the original GAN
In practical training, GAN has some problems, such as difficult training, the loss of generator and discriminator can no ...
Added by kristo5747 on Sat, 12 Feb 2022 15:40:29 +0200
pytorch learning record III
1. Neural network: build small actual combat and use Sequential
Take CIFAR10 model as an example
import torch
import torchvision
from tensorboardX import SummaryWriter
from torch import nn
from torch.nn import ReLU, Sigmoid, Linear, Conv2d, MaxPool2d, Flatten
from torch.utils.data import DataLoader
class test_cifar(nn.Module):
def _ ...
Added by jmcall10 on Sat, 12 Feb 2022 02:04:20 +0200
Word2vec (skip gram and CBOW) - PyTorch
Word vector is a vector used to express the meaning of words, and can also be regarded as the feature vector of words. The technology of mapping words to real vectors is called word embedding.
1, Word embedding (Word2vec)
The unique heat vector can not accurately express the similarity between different words. word2vec is proposed to so ...
Added by j0n on Fri, 11 Feb 2022 13:14:28 +0200
8 Python libraries that can improve the efficiency of data science and save valuable time
In data science, you may waste a lot of time coding and waiting for the computer to run something. So I chose some Python libraries that can help you save valuable time.1,OptunaOptuna is an open source hyperparametric optimization framework, which can automatically find the best hyperparameters for machine learning models.The most basic (and po ...
Added by Horatiu on Fri, 11 Feb 2022 08:29:27 +0200
The underlying storage mode of tensor in Python, dimension transformation, permute/view/reshape, dimension size and number
Record the underlying storage mode of tensor in pytorch, dimension transformation, permute/view/reshape, dimension size and number.
The underlying storage mode of tensor
The underlying storage of tensor is based on the principle of row priority, such as:
>>import torch
>>a=tensor.rand((2,2,3))
>>a
tensor([[[0.1345,0.4907,0. ...
Added by racerxfactor on Thu, 10 Feb 2022 09:54:12 +0200
Attention is All You Need paper notes and pytorch code Notes
Self reference Li Mu read the paper and pytorch code
I don't understand
residual networkPosition-wiseLayer normEncoder attention
Parameter setting
## dimension
d_model = 512 # Dimensions of sub layers, embedding layers and outputs (an addition operation to make use of residual connection)
d_inner_hid = 2048 # Dimension of Feed Forward(MLP) ...
Added by lyasian on Wed, 09 Feb 2022 17:29:29 +0200