text stringlengths 0 93.6k |
|---|
# <FILESEP> |
import os |
USED_DEVICES = "-1" # if your want to use CPU in a server with GPU, change "0" to "-1" |
os.environ["CUDA_DEVICE_ORDER"] = "PCI_BUS_ID" |
os.environ["CUDA_VISIBLE_DEVICES"] = USED_DEVICES |
os.environ["CUDA_LAUNCH_BLOCKING"] = "1" |
import alphastarmini |
import torch |
from alphastarmini.core.arch import entity_encoder |
from alphastarmini.core.arch import scalar_encoder |
from alphastarmini.core.arch import spatial_encoder |
from alphastarmini.core.arch import arch_model |
from alphastarmini.core.arch import action_type_head |
from alphastarmini.core.arch import selected_units_head |
from alphastarmini.core.arch import target_unit_head |
from alphastarmini.core.arch import delay_head |
from alphastarmini.core.arch import queue_head |
from alphastarmini.core.arch import location_head |
from alphastarmini.core.arch import agent |
from alphastarmini.core.arch import baseline |
from alphastarmini.core.sl import load_pickle |
from alphastarmini.core.rl import action |
from alphastarmini.core.rl import env_utils |
from alphastarmini.core.rl import actor |
from alphastarmini.core.rl import against_computer |
from alphastarmini.core.rl import pseudo_reward |
import param as P |
if __name__ == '__main__': |
# if we don't add this line, it may cause running time error while in Windows |
# torch.multiprocessing.freeze_support() |
print("run init") |
# ------------------------ |
# 1. we transform the replays to pickle |
from alphastarmini.core.sl import transform_replay_data |
transform_replay_data.test(on_server=P.on_server) |
print('run over') |
# <FILESEP> |
import json |
import numpy as np |
import torch |
import os |
import warnings |
import wandb |
import random |
from torch.utils.data import Sampler, Dataset |
import matplotlib.pyplot as plt |
import matplotlib |
import PIL |
# used to sample a subset of the val set during training |
class SubsetSampler(Sampler): |
def __init__(self, mask): |
self.mask = mask |
def __iter__(self): |
return (self.indices[i] for i in torch.nonzero(self.mask)) |
def __len__(self): |
return len(self.mask) |
class ListDataset(Dataset): |
def __init__(self, original_list): |
self.original_list = original_list |
def __len__(self): |
return len(self.original_list) |
def __getitem__(self, i): |
return self.original_list[i] |
def load_config(args): |
if args.eval != -1: |
path = f'./configs/eval_ssm_config_{args.eval}.json' |
else: |
path = './configs/finetune_ssm_config.json' |
f = open(path) |
json_data = json.load(f) |
f.close() |
if args.device != 'None': |
json_data['model_device'] = f'cuda:{args.device}' |
return json_data |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.