ICode9

精准搜索请尝试: 精确搜索
  • 67自注意力和位置编码2022-08-17 23:02:21

    点击查看代码 import math import torch from torch import nn from d2l import torch as d2l # 自注意力 num_hiddens, num_heads = 100, 5 attention = d2l.MultiHeadAttention(num_hiddens, num_hiddens, num_hiddens, num_hiddens, num

  • 68多头注意力2022-08-17 23:01:51

    点击查看代码 import math import torch from torch import nn from d2l import torch as d2l # 选择缩放点积注意力作为每一个注意力头 #

  • 68transformer2022-08-17 23:01:24

    点击查看代码 import math import pandas as pd import torch from torch import nn from d2l import torch as d2l #@save class PositionWiseFFN(nn.Module): """基于位置的前馈网络""" # 全连接 # num_step会变 序列长度 # 所以序列当中每一个元素做一个全连

  • 69BERT2022-08-17 23:00:54

    点击查看代码 import math import pandas as pd import torch from torch import nn from d2l import torch as d2l #@save class PositionWiseFFN(nn.Module): """基于位置的前馈网络""" # 全连接 # num_step会变 序列长度 # 所以序列当中每一个元素做一个全连

  • 57长短期记忆网络LSTM2022-07-04 15:03:30

    点击查看代码 import torch from torch import nn from d2l import torch as d2l batch_size, num_steps = 32, 35 train_iter, vocab = d2l.load_data_time_machine(batch_size, num_steps) # 参数初始化 def get_lstm_params(vocab_size, num_hiddens, device): num_in

  • 56门控循环单元GRU2022-07-04 14:31:48

    点击查看代码 import torch from torch import nn from d2l import torch as d2l """ 额外的控制单元 """ batch_size, num_steps = 32, 35 train_iter, vocab = d2l.load_data_time_machine(batch_size, num_steps) # 初始化模型参数 def get_params(vocab_size

专注分享技术,共同学习,共同进步。侵权联系[81616952@qq.com]

Copyright (C)ICode9.com, All Rights Reserved.

ICode9版权所有