Skip to content

  • 体验新版
    • 正在加载...
  • 登录
  • PaddlePaddle
  • Paddle
  • Issue
  • #2834

P
Paddle
  • 项目概览

PaddlePaddle / Paddle
大约 2 年 前同步成功

通知 2325
Star 20933
Fork 5424
  • 代码
    • 文件
    • 提交
    • 分支
    • Tags
    • 贡献者
    • 分支图
    • Diff
  • Issue 1423
    • 列表
    • 看板
    • 标记
    • 里程碑
  • 合并请求 543
  • Wiki 0
    • Wiki
  • 分析
    • 仓库
    • DevOps
  • 项目成员
  • Pages
P
Paddle
  • 项目概览
    • 项目概览
    • 详情
    • 发布
  • 仓库
    • 仓库
    • 文件
    • 提交
    • 分支
    • 标签
    • 贡献者
    • 分支图
    • 比较
  • Issue 1,423
    • Issue 1,423
    • 列表
    • 看板
    • 标记
    • 里程碑
  • 合并请求 543
    • 合并请求 543
  • Pages
  • 分析
    • 分析
    • 仓库分析
    • DevOps
  • Wiki 0
    • Wiki
  • 成员
    • 成员
  • 收起侧边栏
  • 动态
  • 分支图
  • 创建新Issue
  • 提交
  • Issue看板
已关闭
开放中
Opened 7月 13, 2017 by saxon_zh@saxon_zhGuest

recurrent_group cannot return more than one layer as its outputs.

Created by: lcy-seso

I use machine translation in PaddleBook as the test example.

  • Previously, recurrent_group in PaddlePaddle can return more than one layers inside the layer group as its outputs.
  • The returned layers are accessible to any other layers in the network, while layers inside recurrent_group (and is not returned by recurrent_gropu) are not accessible outside the recurrent_group.
  • I hope to use the above feature, but if I change this line to return more than one layers in recurrent_group, the configuration parsing process crashes. It seems that there is some bug?
  • the error information is as follows:
Traceback (most recent call last):
  File "network_conf.py", line 114, in <module>
    print parse_network(seq_to_seq_net(30000, 30000, True, 3, 250))
  File "network_conf.py", line 108, in seq_to_seq_net
    max_length=max_length)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/v2/config_base.py", line 52, in wrapped
    out = f(*args, **xargs)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/trainer_config_helpers/default_decorators.py", line 53, in __wrapper__
    return func(*args, **kwargs)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/trainer_config_helpers/layers.py", line 3949, in beam_search
    is_generating=True)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/trainer_config_helpers/default_decorators.py", line 53, in __wrapper__
    return func(*args, **kwargs)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/trainer_config_helpers/layers.py", line 3657, in recurrent_group
    RecurrentLayerGroupEnd(name=name)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/trainer/config_parser.py", line 396, in RecurrentLayerGroupEnd
    "memory declare wrong size:%d" % memory_link.size)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/trainer/config_parser.py", line 156, in config_assert
    logger.fatal(msg)
  File "/home/caoying/paddle_codes/models/nmt_plot_attention_weights/paddle/trainer/config_parser.py", line 3531, in my_fatal
    raise Exception()
Exception
指派人
分配到
无
里程碑
无
分配里程碑
工时统计
无
截止日期
无
标识: paddlepaddle/Paddle#2834
渝ICP备2023009037号

京公网安备11010502055752号

网络110报警服务 Powered by GitLab CE v13.7
开源知识
Git 入门 Pro Git 电子书 在线学 Git
Markdown 基础入门 IT 技术知识开源图谱
帮助
使用手册 反馈建议 博客
《GitCode 隐私声明》 《GitCode 服务条款》 关于GitCode
Powered by GitLab CE v13.7