Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bug fix and refine cloud training for DS2 #201

Merged
merged 3 commits into from
Aug 14, 2017

Conversation

xinghai-sun
Copy link
Contributor

Resolve #200

  • Add missing is_local argument (when set False, use pserver).
  • Add exception thrown if cloud cp failed.
  • Add cloud mkdir if cloud path for uploading does not exist.
  • Fix a bug using common path ./local_manifest for all nodes. (convert to /local_manifest)
  • Refine coding style.

Summary:
1. Add missing is_local argument (when set False, use pserver).
2. Add exception thrown if cp failed.
3. Add cloud mkdir if not cloud path for uploading does not exist.
4. Fix a bug using common path ./local_manifest for all nodes. (convert to /local_manifest)
5. Refine coding style.
Copy link
Contributor

@wanghaoshuang wanghaoshuang left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM.

CLOUD_DATA_DIR="/pfs/dlnel/home/sunxinghai@baidu.com/deepspeech2/data"
CLOUD_MODEL_DIR="/pfs/dlnel/home/sunxinghai@baidu.com/deepspeech2/model"
# Configure cloud resources
NUM_CPU=12
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

CPU和GPU应该设置一样,真实分配的GPU数量是来自--cpu,这是PaddleCloud的bug.
这里如果写成12,有可能会调度困难或报错。

@xinghai-sun xinghai-sun merged commit 406db45 into PaddlePaddle:develop Aug 14, 2017
@xinghai-sun xinghai-sun deleted the refine_cloud branch August 14, 2017 14:02
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants