diff --git a/cv/point_cloud/Point-BERT/pytorch/prepare.sh b/cv/point_cloud/Point-BERT/pytorch/prepare.sh index ed751a09e7766b4feb21ff8dbd91cec6f4c36b09..cabec97ee064ec58393056639dbabddca253fdfa 100644 --- a/cv/point_cloud/Point-BERT/pytorch/prepare.sh +++ b/cv/point_cloud/Point-BERT/pytorch/prepare.sh @@ -31,7 +31,7 @@ pip3 install --upgrade https://github.com/unlimblue/KNN_CUDA/releases/download/0 cd data/ShapeNet55-34 if [[ ! -f "ShapeNet55.zip" ]]; then - wget http://10.150.9.95/swapp/datasets/cv/ShapeNet55.zip + wget http://files.deepspark.org.cn:880/deepspark/ShapeNet55.zip unzip ShapeNet55.zip mv ShapeNet55/shapenet_pc/ . rm -r ShapeNet55 @@ -42,7 +42,7 @@ cd - cd ./data/ModelNet/modelnet40_normal_resampled if [[ ! -f "processed_ModelNet.zip" ]]; then - wget http://10.150.9.95/swapp/datasets/cv/processed_ModelNet.zip + wget http://files.deepspark.org.cn:880/deepspark/processed_ModelNet.zip unzip processed_ModelNet.zip mv processed_ModelNet/* . fi diff --git a/nlp/language_model/bert/mindspore/scripts/run_squad_gpu.sh b/nlp/language_model/bert/mindspore/scripts/run_squad_gpu.sh index 93490eb17532001726ac1a49f3535d0b732fbf76..65623776f5fbf466effe0a502e2d9a31c23ce296 100644 --- a/nlp/language_model/bert/mindspore/scripts/run_squad_gpu.sh +++ b/nlp/language_model/bert/mindspore/scripts/run_squad_gpu.sh @@ -34,7 +34,7 @@ mkdir -p squad training_tf_record_file=squad/train.tf_record if [[ ! -f "${training_tf_record_file}" ]]; then cd squad - wget http://10.201.40.22/files/datasets/bert/squad_data_with_tf_record.tar + wget http://files.deepspark.org.cn:880/deepspark/squad_data_with_tf_record.tar tar xvf squad_data_with_tf_record.tar rm -rf squad_data_with_tf_record.tar cd - diff --git a/toolbox/Megatron-DeepSpeed/dataset/convert_dahoas.sh b/toolbox/Megatron-DeepSpeed/dataset/convert_dahoas.sh index 07a3fe50ba2a52bef09f45634544883170368f51..1405318cd6a8fdf967859e686b84a0c0fde0734f 100644 --- a/toolbox/Megatron-DeepSpeed/dataset/convert_dahoas.sh +++ b/toolbox/Megatron-DeepSpeed/dataset/convert_dahoas.sh @@ -2,7 +2,7 @@ CUR_DIR=$(cd "$(dirname "$0")";pwd) if [[ ! -e ${CUR_DIR}/dahoas_train.jsonl ]]; then - wget http://sw.iluvatar.ai/download/apps/datasets/nlp/RLHF/dahoas_train.jsonl + wget http://files.deepspark.org.cn:880/deepspark/dahoas_train.jsonl fi PROJ_HOME=$(dirname "$PWD") diff --git a/toolbox/Megatron-DeepSpeed/dataset/download_and_covert_llama3_dataset.sh b/toolbox/Megatron-DeepSpeed/dataset/download_and_covert_llama3_dataset.sh index 432d6d9b0736ffa52a8d4804123e1a94f5db67b5..25981c00fdbb0d993297b8fe90bd0b89de489e40 100644 --- a/toolbox/Megatron-DeepSpeed/dataset/download_and_covert_llama3_dataset.sh +++ b/toolbox/Megatron-DeepSpeed/dataset/download_and_covert_llama3_dataset.sh @@ -3,7 +3,7 @@ set -euox pipefail CUR_DIR=$(pwd) if [[ ! -f $CUR_DIR/small-117M.train.jsonl ]]; then - wget http://10.150.9.95/swapp/datasets/nlp/gpt-2-output-dataset/small-117M.train.jsonl + wget http://files.deepspark.org.cn:880/deepspark/small-117M.train.jsonl fi PROJ_HOME=$(dirname "$PWD") diff --git a/toolbox/Megatron-DeepSpeed/dataset/download_llama2_gpt_small.sh b/toolbox/Megatron-DeepSpeed/dataset/download_llama2_gpt_small.sh index 5141346d0d2b2a8503e7850dd6fa86c1ac6543f9..464359d83eac9979d5990d42ffc2c24e242890f8 100644 --- a/toolbox/Megatron-DeepSpeed/dataset/download_llama2_gpt_small.sh +++ b/toolbox/Megatron-DeepSpeed/dataset/download_llama2_gpt_small.sh @@ -6,6 +6,6 @@ cd ${CUR_DIR} # llama2 if [[ ! -d ${CUR_DIR}/gpt_small_117M ]]; then echo "gpt_small dataset not exist, downloading..." - wget http://10.150.9.95/swapp/datasets/nlp/gpt-2-output-dataset/gpt_small_117M.tar + wget http://files.deepspark.org.cn:880/deepspark/gpt_small_117M.tar tar -xf gpt_small_117M.tar && rm -f gpt_small_117M.tar fi diff --git a/toolbox/Megatron-DeepSpeed/dataset/download_llama3_gpt_small.sh b/toolbox/Megatron-DeepSpeed/dataset/download_llama3_gpt_small.sh index f7e6e8358d244e16af3272b18b9ab058ac7b9b58..995a69a07973b857817449f96aaeff4025d96a17 100644 --- a/toolbox/Megatron-DeepSpeed/dataset/download_llama3_gpt_small.sh +++ b/toolbox/Megatron-DeepSpeed/dataset/download_llama3_gpt_small.sh @@ -6,6 +6,6 @@ cd ${CUR_DIR} ## llama3 if [[ ! -d ${CUR_DIR}/gpt_small_117M_llama3 ]]; then echo "gpt_small dataset not exist, downloading..." - wget http://10.150.9.95/swapp/datasets/nlp/gpt-2-output-dataset/gpt_small_117M_llama3.tar + wget http://files.deepspark.org.cn:880/deepspark/gpt_small_117M_llama3.tar tar -xf gpt_small_117M_llama3.tar && rm -f gpt_small_117M_llama3.tar fi \ No newline at end of file