109 Star 878 Fork 1.5K

MindSpore/models

加入 Gitee
与超过 1200万 开发者一起发现、参与优秀开源项目,私有仓库也完全免费 :)
免费加入
文件
克隆/下载
贡献代码
同步代码
取消
提示: 由于 Git 不支持空文件夾,创建文件夹后会生成空的 .keep 文件
Loading...
README

GOMO Example

Description

Generalized Operator Modelling of the Ocean (GOMO) is a three-dimensional ocean model based on OpenArray which is a simple operator library for the decoupling of ocean modelling and parallel computing (Xiaomeng Huang et al, 2019). GOMO is a numerical solution model using finite differential algorithm to solve PDE equations. With MindSpore and GPU, we can achieve great improvements in solving those PDE equations compared with CPU. This is an example of training GOMO Model with MindSpore on GPU.

Model Architecture

The overall model architecture of GOMO is show below:link. The fundamental equations and algorithms of GOMO can also be found in this article

Dataset

Dataset used: Seamount

  • Dataset size: 65x49x21

  • Data format:nc

  • Download the dataset

download the GOMO(https://github.com/hxmhuang/GOMO) from Github and you can find the seamount dataset file in the GOMO/bin/data directory.

Environment Requirements

Quick Start

After installing MindSpore via the official website, you can start training as follows:

# run distributed training example
bash run_distribute_train.sh [im] [jm] [kb] [step] [DATASET_PATH]

Script Description

Script and Sample Code

└── ocean_model
    ├── README.md                                 # descriptions about ocean model GOMO
    ├── scripts
    │    ├── run_distribute_train.sh               # launch distributed training for GPU
    ├──src
    │    ├── GOMO.py                               # GOMO model
    │    ├── Grid.py                               # grid initial
    │    ├── stencil.py                            # averaging and differential stencil oprator
    │    ├── op_operator.py                        # averaging and differential kernel operator
    │    ├── read_var.py                           # read variables from nc file
    ├── train.py                                  # train script

Training Process

bash run_distribute_train.sh [im] [jm] [kb] [step] [DATASET_PATH]

Training result will be stored in the current path, whose folder name begins with "train".

Model Description

Evaluation Performance

Parameters GPU
Resource GPU(Tesla V100 SXM2),Memory 16G
uploaded Date
MindSpore Version
Dataset Seamount
Training Parameters step=10, im=65, km=49, kb=21
Outputs numpy file
Speed 17 ms/step
Total time 3 mins
Scripts Link

Description of Random Situation

ModelZoo HomePage

Please check the official homepage.

马建仓 AI 助手
尝试更多
代码解读
代码找茬
代码优化
1
https://gitee.com/mindspore/models.git
git@gitee.com:mindspore/models.git
mindspore
models
models
master

搜索帮助