From 290f5515b71edb1a7f342c7a298a8b0c94de34b9 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Mon, 25 Apr 2022 16:23:25 +0700 Subject: [PATCH 01/17] update scripts --- Assets/Scripts/Bots/CharacterFactory.cs | 18 +++++---- .../Scripts/Character/CharacterCondition.cs | 17 +++++++-- .../Scripts/Character/MovementController.cs | 21 ++++++++-- Assets/Scripts/Character/NPC.cs | 27 ++++++++----- Assets/Scripts/Managers/GameManager.cs | 14 +++++++ Assets/Scripts/Misc/FlagZone.cs | 6 +-- Assets/Scripts/Misc/Settings.cs | 38 ++++++++++--------- Assets/Scripts/Pickups/AmmoPickUp.cs | 2 +- Assets/Scripts/Pickups/ArmourPickUp.cs | 2 +- Assets/Scripts/Pickups/HealthPickUp.cs | 2 +- Assets/Scripts/Pickups/PickUpSpawner.cs | 2 +- Assets/Scripts/Utils/BoolToInteger.cs | 7 ++++ Assets/Scripts/Utils/BoolToInteger.cs.meta | 11 ++++++ 13 files changed, 120 insertions(+), 47 deletions(-) create mode 100644 Assets/Scripts/Utils/BoolToInteger.cs create mode 100644 Assets/Scripts/Utils/BoolToInteger.cs.meta diff --git a/Assets/Scripts/Bots/CharacterFactory.cs b/Assets/Scripts/Bots/CharacterFactory.cs index e2c8be3..ab18289 100644 --- a/Assets/Scripts/Bots/CharacterFactory.cs +++ b/Assets/Scripts/Bots/CharacterFactory.cs @@ -25,10 +25,10 @@ public class CharacterFactory : MonoBehaviour private void Start() { - var attcNum = SettingsReader.Instance.GetSettings.numOfAttackers; - var defNum = SettingsReader.Instance.GetSettings.numOfDefenders; - var humanDef = SettingsReader.Instance.GetSettings.hasHumanDefender == true ? 1 : 0; - var humanAtc = SettingsReader.Instance.GetSettings.hasHumanAttacker == true ? 1 : 0; + var attcNum = SettingsReader.Instance.GetSettings.NumOfAttackers; + var defNum = SettingsReader.Instance.GetSettings.NumOfDefenders; + var humanDef = SettingsReader.Instance.GetSettings.HasHumanDefender == true ? 1 : 0; + var humanAtc = SettingsReader.Instance.GetSettings.HasHumanAttacker == true ? 1 : 0; if (humanAtc == 1 && humanDef == 1) throw new System.ArgumentException("Can be only one human player"); @@ -56,6 +56,10 @@ public class CharacterFactory : MonoBehaviour spawnPoint.position, Quaternion.identity); gameobject.SetActive(true); + if (team == Team.Attackers) + gameObject.tag = "Attacker"; + else + gameObject.tag = "Defender"; if (typeAi == TypeAI.HumanAI) { @@ -65,7 +69,7 @@ public class CharacterFactory : MonoBehaviour else { gameobject.GetComponent().GetCharacter.Team = team; - gameobject.GetComponent().currentPosition = spawnPoint; + gameobject.GetComponent().CurrentNavPoint = spawnPoint; Bots.Add(gameobject); } } @@ -81,8 +85,8 @@ public class CharacterFactory : MonoBehaviour else bot.transform.position = spawnPointsForDefendersTeam[Random.Range(0, spawnPointsForDefendersTeam.Count)].position; } - var player = Player.GetComponent(); - if (player != null) + Player player; + if (TryGetComponent(out player)) { player.ResetCharacter(); if (player.GetCharacter.Team == Team.Attackers) diff --git a/Assets/Scripts/Character/CharacterCondition.cs b/Assets/Scripts/Character/CharacterCondition.cs index d1ec569..b674222 100755 --- a/Assets/Scripts/Character/CharacterCondition.cs +++ b/Assets/Scripts/Character/CharacterCondition.cs @@ -21,6 +21,17 @@ public class CharacterCondition OnChangeHealthEvent?.Invoke(value); } } + + public int GetHealthPointsInQuantile() + { + if (health < 25) + return 0; + else if (health < 50) + return 1; + else if (health < 75) + return 2; + else return 3; + } private int armour; public int ArmourPoints { @@ -51,9 +62,9 @@ public class CharacterCondition public CharacterCondition() { var settings = SettingsReader.Instance.GetSettings; - ammo = settings.maxAmmo; - health = settings.maxHealth; - armour = settings.maxArmour; + ammo = settings.MaxAmmo; + health = settings.MaxHealth; + armour = settings.MaxArmour; } public void GiveHealth(int health) => HealthPoints = Mathf.Clamp(health + HealthPoints, 0, 100); diff --git a/Assets/Scripts/Character/MovementController.cs b/Assets/Scripts/Character/MovementController.cs index 20e1026..334b7d2 100644 --- a/Assets/Scripts/Character/MovementController.cs +++ b/Assets/Scripts/Character/MovementController.cs @@ -2,16 +2,31 @@ using System.Collections.Generic; using UnityEngine; using UnityEngine.AI; +using System.Threading.Tasks; [RequireComponent(typeof(NavMeshAgent))] public class MovementController : MonoBehaviour { - public NavPoint currentPosition { get; set; } + public NavPoint CurrentNavPoint { get; set; } + public float FlagDistance { get; private set; } + private GameObject flag; + private const float updateFlagPositionDelay = 5; [SerializeField] private NavMeshAgent navMeshAgent; private void Start() { - navMeshAgent.speed = SettingsReader.Instance.GetSettings.movementSpeed; + navMeshAgent.speed = SettingsReader.Instance.GetSettings.MovementSpeed; + InvokeRepeating(nameof(UpdateFlagPosition), 0, updateFlagPositionDelay); + } + + private void OnDestroy() + { + CancelInvoke(nameof(UpdateFlagPosition)); + } + + private void UpdateFlagPosition() + { + FlagDistance = (flag.transform.position - gameObject.transform.position).magnitude; } public void MoveToRandomPoint() @@ -23,7 +38,7 @@ public class MovementController : MonoBehaviour public List getPointsCandidate() { return MapManager.navPoints - .Where(point => (currentPosition.position - point.position).magnitude < SettingsReader.Instance.GetSettings.movementSpeed) + .Where(point => (CurrentNavPoint.position - point.position).magnitude < SettingsReader.Instance.GetSettings.MovementSpeed) .ToList(); } diff --git a/Assets/Scripts/Character/NPC.cs b/Assets/Scripts/Character/NPC.cs index 1069d42..c738d0b 100644 --- a/Assets/Scripts/Character/NPC.cs +++ b/Assets/Scripts/Character/NPC.cs @@ -3,7 +3,6 @@ using UnityEngine; using Unity.MLAgents; using Unity.MLAgents.Sensors; using Unity.MLAgents.Actuators; -using System.Collections.Generic; [RequireComponent(typeof(MovementController))] public class NPC : Agent, ICharacter @@ -11,6 +10,7 @@ public class NPC : Agent, ICharacter [HideInInspector] public Character AgentCharacter; public CharacterCondition Condition; + private FlagZone flagZone; public NPC_BaseState NPC_State { get; private set; } @@ -47,29 +47,38 @@ public class NPC : Agent, ICharacter public override void OnEpisodeBegin() { NPC_State = DirectState; + flagZone = GameObject.FindObjectOfType(); } public override void CollectObservations(VectorSensor sensor) { + var candidates = moveController.getPointsCandidate(); + sensor.AddObservation(Condition.HealthPoints); sensor.AddObservation(Condition.ArmourPoints); sensor.AddObservation(Condition.Ammunition); sensor.AddObservation((int)NPC_State.State); - - var candidates = moveController.getPointsCandidate(); + sensor.AddObservation((!flagZone.isNotOccup).ToInt()); + sensor.AddObservation(AgentCharacter.LastTimeHit); + sensor.AddObservation(Condition.GetHealthPointsInQuantile()); + sensor.AddObservation(candidates.Count); + sensor.AddObservation(GameManager.IsEnemyNearby(gameObject.transform.position, AgentCharacter.Team)); + foreach (var point in candidates) { + Debug.Log((float)moveController.CurrentNavPoint.PointId); + bufferSensor.AppendObservation(new float[] { //1 position in navpointId - (float)moveController.currentPosition.PointId, + (float)moveController.CurrentNavPoint.PointId, //2 distance to flag - moveController.currentPosition.FlagDistance, + moveController.FlagDistance, //3 death count in point - moveController.currentPosition.DeathAttr, + moveController.CurrentNavPoint.DeathAttr, //4 flagEnemyDistance - GameManager.IsCloserToFlagFromNextNavPoint(point, transform.position)==true?1:0, + GameManager.IsCloserToFlagFromNextNavPoint(point, transform.position).ToInt(), //5 EnemyVsNavPointDistance - GameManager.IsCloserToEnemyThanToNextNavPoint(point,transform.position, AgentCharacter.Team)==true?1:0 + GameManager.IsCloserToEnemyThanToNextNavPoint(point,transform.position, AgentCharacter.Team).ToInt() }); } @@ -103,7 +112,7 @@ public class NPC : Agent, ICharacter if (Condition.HealthPoints < 0) { OnKilledEvent?.Invoke(this); - moveController.currentPosition.DeathAttr += 1; + moveController.CurrentNavPoint.DeathAttr += 1; } } diff --git a/Assets/Scripts/Managers/GameManager.cs b/Assets/Scripts/Managers/GameManager.cs index 54d1e1d..b91908d 100755 --- a/Assets/Scripts/Managers/GameManager.cs +++ b/Assets/Scripts/Managers/GameManager.cs @@ -51,6 +51,20 @@ public class GameManager : MonoBehaviour return false; } + public static bool IsEnemyNearby(Vector3 currentTransform, Team team) + { + SimpleMultiAgentGroup agentGroup; + if (team == Team.Attackers) + agentGroup = AttackersTeam; + else + agentGroup = DefendersTeam; + + foreach (var agent in agentGroup.GetRegisteredAgents()) + if ((currentTransform - agent.transform.position).magnitude < SettingsReader.Instance.GetSettings.ViewDistance) + return true; + return false; + } + public static bool IsCloserToFlagFromNextNavPoint(NavPoint navPoint, Vector3 currentTransform) => navPoint.FlagDistance < (currentTransform - GameObject.FindGameObjectWithTag("Flag").transform.position).magnitude; diff --git a/Assets/Scripts/Misc/FlagZone.cs b/Assets/Scripts/Misc/FlagZone.cs index 8cd4ab4..c6328dd 100755 --- a/Assets/Scripts/Misc/FlagZone.cs +++ b/Assets/Scripts/Misc/FlagZone.cs @@ -9,14 +9,14 @@ public class FlagZone : MonoBehaviour public float TimeStayDefenders { get; private set; } private int occupDefenders; private int occupAttackers; - private bool isOccupBoth => (occupDefenders>0) && (occupAttackers>0); - private bool isNotOccup => (occupDefenders == 0) && (occupAttackers == 0); + public bool isOccupBoth => (occupDefenders>0) && (occupAttackers>0); + public bool isNotOccup => (occupDefenders == 0) && (occupAttackers == 0); private float timeForWin; private void Start() { - timeForWin = SettingsReader.Instance.GetSettings.timeToWin; + timeForWin = SettingsReader.Instance.GetSettings.TimeToWin; TimeStayAttackers = 0; TimeStayDefenders = 0; occupAttackers = 0; diff --git a/Assets/Scripts/Misc/Settings.cs b/Assets/Scripts/Misc/Settings.cs index e01abc1..b1b7ea8 100755 --- a/Assets/Scripts/Misc/Settings.cs +++ b/Assets/Scripts/Misc/Settings.cs @@ -3,30 +3,32 @@ [CreateAssetMenu(fileName ="Game Settings", menuName = "Game/Settings", order = 51)] public class Settings : ScriptableObject { - public bool isTesting; + public bool IsTesting; - public float timeToWin; - public float timeOut; + public float TimeToWin; + public float TimeOut; [Header("movement")] - public float movementDistance; - public float movementSpeed; + public float MovementDistance; + public float MovementSpeed; - public TypeAI defTeamAI; - public TypeAI atcTeamAI; - public int numOfDefenders; - public int numOfAttackers; - public bool hasHumanDefender; - public bool hasHumanAttacker; + public TypeAI DefTeamAI; + public TypeAI AtcTeamAI; + public int NumOfDefenders; + public int NumOfAttackers; + public bool HasHumanDefender; + public bool HasHumanAttacker; - public int healthPickupAmount; - public int armourPickupAmount; - public int ammunitionPickupAmount; - public int pickupsAmount; + public int HealthPickupAmount; + public int ArmourPickupAmount; + public int AmmunitionPickupAmount; + public int PickupsAmount; - public int maxHealth; - public int maxArmour; - public int maxAmmo; + public int MaxHealth; + public int MaxArmour; + public int MaxAmmo; + + public float ViewDistance; public float GetHitChanceInDirectPoint; public float GetHitChanceInRunning; diff --git a/Assets/Scripts/Pickups/AmmoPickUp.cs b/Assets/Scripts/Pickups/AmmoPickUp.cs index 506598d..8439c21 100755 --- a/Assets/Scripts/Pickups/AmmoPickUp.cs +++ b/Assets/Scripts/Pickups/AmmoPickUp.cs @@ -13,7 +13,7 @@ public class AmmoPickUp : MonoBehaviour, IPickable public void PickObject(GameObject obj) { - obj.GetComponent()?.GetCharacter.Condition.TakeAmmo(SettingsReader.Instance.GetSettings.ammunitionPickupAmount); + obj.GetComponent()?.GetCharacter.Condition.TakeAmmo(SettingsReader.Instance.GetSettings.AmmunitionPickupAmount); gameObject.SetActive(false); } } diff --git a/Assets/Scripts/Pickups/ArmourPickUp.cs b/Assets/Scripts/Pickups/ArmourPickUp.cs index 86be289..c68e491 100755 --- a/Assets/Scripts/Pickups/ArmourPickUp.cs +++ b/Assets/Scripts/Pickups/ArmourPickUp.cs @@ -13,7 +13,7 @@ public class ArmourPickUp : MonoBehaviour, IPickable public void PickObject(GameObject obj) { - obj.GetComponent()?.GetCharacter.Condition.GiveArmour(SettingsReader.Instance.GetSettings.armourPickupAmount); + obj.GetComponent()?.GetCharacter.Condition.GiveArmour(SettingsReader.Instance.GetSettings.ArmourPickupAmount); gameObject.SetActive(false); } } diff --git a/Assets/Scripts/Pickups/HealthPickUp.cs b/Assets/Scripts/Pickups/HealthPickUp.cs index f92345a..36f81ba 100755 --- a/Assets/Scripts/Pickups/HealthPickUp.cs +++ b/Assets/Scripts/Pickups/HealthPickUp.cs @@ -13,7 +13,7 @@ public class HealthPickUp : MonoBehaviour, IPickable public void PickObject(GameObject obj) { - obj.GetComponent()?.GetCharacter.Condition.GiveHealth(SettingsReader.Instance.GetSettings.healthPickupAmount); + obj.GetComponent()?.GetCharacter.Condition.GiveHealth(SettingsReader.Instance.GetSettings.HealthPickupAmount); gameObject.SetActive(false); } } diff --git a/Assets/Scripts/Pickups/PickUpSpawner.cs b/Assets/Scripts/Pickups/PickUpSpawner.cs index 945c1d3..7a29705 100755 --- a/Assets/Scripts/Pickups/PickUpSpawner.cs +++ b/Assets/Scripts/Pickups/PickUpSpawner.cs @@ -27,7 +27,7 @@ public class PickUpSpawner : MonoBehaviour private void Start() { pickups = new List(); - var amount = SettingsReader.Instance.GetSettings.pickupsAmount; + var amount = SettingsReader.Instance.GetSettings.PickupsAmount; for (int i = 0; i < amount; i++) pickups.Add(GameObject.Instantiate(healthPrefab, spawnPoints[Random.Range(0, spawnPoints.Count)].transform.position, Quaternion.identity)); for (int i = 0; i < amount; i++) diff --git a/Assets/Scripts/Utils/BoolToInteger.cs b/Assets/Scripts/Utils/BoolToInteger.cs new file mode 100644 index 0000000..0d8090a --- /dev/null +++ b/Assets/Scripts/Utils/BoolToInteger.cs @@ -0,0 +1,7 @@ +public static class BoolExtension +{ + public static int ToInt(this bool _bool) + { + return _bool == true ? 1 : 0; + } +} \ No newline at end of file diff --git a/Assets/Scripts/Utils/BoolToInteger.cs.meta b/Assets/Scripts/Utils/BoolToInteger.cs.meta new file mode 100644 index 0000000..3688775 --- /dev/null +++ b/Assets/Scripts/Utils/BoolToInteger.cs.meta @@ -0,0 +1,11 @@ +fileFormatVersion: 2 +guid: f48fff3c2eda14d4fba923fe8875f651 +MonoImporter: + externalObjects: {} + serializedVersion: 2 + defaultReferences: [] + executionOrder: 0 + icon: {instanceID: 0} + userData: + assetBundleName: + assetBundleVariant: -- 2.49.0 From c8af0e528415cb8b3b198edc7eb513fdf9133710 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Wed, 4 May 2022 23:50:07 +0700 Subject: [PATCH 02/17] to new git --- Assets/Scripts/Bots/CharacterFactory.cs | 44 +++-- Assets/Scripts/Bots/TeamEnum.cs | 11 ++ Assets/Scripts/Character/Character.cs | 6 - .../Scripts/Character/CharacterCondition.cs | 32 +++- .../Interfaces.meta} | 2 +- .../Character/Interfaces/ICharacter.cs | 5 + .../Interfaces/ICharacter.cs.meta} | 2 +- .../Character/Interfaces/INpcBaseState.cs | 17 ++ .../Interfaces/INpcBaseState.cs.meta} | 2 +- .../Scripts/Character/MovementController.cs | 70 +++++-- Assets/Scripts/Character/NPC.cs | 178 ++++++++++++------ Assets/Scripts/Character/NPC_State.cs | 46 ----- Assets/Scripts/Character/NpcState.cs | 68 +++++++ .../{NPC_State.cs.meta => NpcState.cs.meta} | 0 Assets/Scripts/Character/Player.cs | 12 +- .../Character/scr_CharacterController.cs | 64 +++---- Assets/Scripts/Character/scr_Models.cs | 120 ++++++------ Assets/Scripts/Managers/GameManager.cs | 107 ++++++++--- Assets/Scripts/Managers/MapManager.cs | 63 ++++++- Assets/Scripts/Managers/TimeManager.cs | 8 +- Assets/Scripts/Misc/FlagZone.cs | 14 +- Assets/Scripts/Misc/NavPoint.cs | 20 +- Assets/Scripts/Misc/Settings.cs | 4 +- Assets/Scripts/Misc/SettingsReader.cs | 16 +- Assets/Scripts/Misc/Statistics.cs | 9 - Assets/Scripts/Pickups/AmmoPickUp.cs | 8 +- Assets/Scripts/Pickups/ArmourPickUp.cs | 8 +- Assets/Scripts/Pickups/HealthPickUp.cs | 8 +- Assets/Scripts/Pickups/IPickable.cs | 5 +- Assets/Scripts/Pickups/PickUpSpawner.cs | 10 +- Assets/Scripts/Sensors/SensorType.cs | 6 - Assets/Scripts/Sensors/Sensors.cs | 4 - Assets/Scripts/Statistics.meta | 8 + Assets/Scripts/Statistics/Logger.cs | 19 ++ Assets/Scripts/Statistics/Logger.cs.meta | 11 ++ Assets/Scripts/Statistics/StatisticManager.cs | 51 +++++ .../StatisticManager.cs.meta} | 2 +- .../Scripts/Utils/SerializableDictionary.cs | 4 +- .../Scripts/Weapons/scr_WeaponController.cs | 15 +- 39 files changed, 720 insertions(+), 359 deletions(-) rename Assets/Scripts/{Sensors.meta => Character/Interfaces.meta} (77%) mode change 100755 => 100644 create mode 100644 Assets/Scripts/Character/Interfaces/ICharacter.cs rename Assets/Scripts/{Sensors/SensorType.cs.meta => Character/Interfaces/ICharacter.cs.meta} (83%) mode change 100755 => 100644 create mode 100644 Assets/Scripts/Character/Interfaces/INpcBaseState.cs rename Assets/Scripts/{Sensors/Sensors.cs.meta => Character/Interfaces/INpcBaseState.cs.meta} (83%) delete mode 100644 Assets/Scripts/Character/NPC_State.cs create mode 100644 Assets/Scripts/Character/NpcState.cs rename Assets/Scripts/Character/{NPC_State.cs.meta => NpcState.cs.meta} (100%) delete mode 100755 Assets/Scripts/Misc/Statistics.cs delete mode 100755 Assets/Scripts/Sensors/SensorType.cs delete mode 100644 Assets/Scripts/Sensors/Sensors.cs create mode 100644 Assets/Scripts/Statistics.meta create mode 100644 Assets/Scripts/Statistics/Logger.cs create mode 100644 Assets/Scripts/Statistics/Logger.cs.meta create mode 100644 Assets/Scripts/Statistics/StatisticManager.cs rename Assets/Scripts/{Misc/Statistics.cs.meta => Statistics/StatisticManager.cs.meta} (91%) mode change 100755 => 100644 diff --git a/Assets/Scripts/Bots/CharacterFactory.cs b/Assets/Scripts/Bots/CharacterFactory.cs index ab18289..0ca7c1d 100644 --- a/Assets/Scripts/Bots/CharacterFactory.cs +++ b/Assets/Scripts/Bots/CharacterFactory.cs @@ -1,26 +1,28 @@ using System.Collections.Generic; using UnityEngine; -using Unity; public class CharacterFactory : MonoBehaviour { - private CharacterFactory instance; - public CharacterFactory Instance { get { return instance; } } + private static CharacterFactory instance; + public static CharacterFactory Instance => instance; [SerializeField] private List spawnPointsForDefendersTeam; [SerializeField] private List spawnPointsForAttackersTeam; [SerializeField] private GameObject AIPrefab; [SerializeField] private GameObject PlayerPrefab; - private List Bots = new List(); - private GameObject Player; + private List bots = new List(); + public GameObject player { get; private set; } private void Awake() { if (instance == null) instance = this; else + { Destroy(gameObject); + Debug.LogError("Only 1 Instance"); + } } private void Start() @@ -53,7 +55,7 @@ public class CharacterFactory : MonoBehaviour { var gameobject = GameObject.Instantiate( typeAi == TypeAI.HumanAI ? PlayerPrefab : AIPrefab, - spawnPoint.position, + spawnPoint.Position, Quaternion.identity); gameobject.SetActive(true); if (team == Team.Attackers) @@ -64,35 +66,49 @@ public class CharacterFactory : MonoBehaviour if (typeAi == TypeAI.HumanAI) { gameobject.GetComponent().GetCharacter.Team = team; - Player = gameobject; + player = gameobject; } else { gameobject.GetComponent().GetCharacter.Team = team; - gameobject.GetComponent().CurrentNavPoint = spawnPoint; - Bots.Add(gameobject); + gameobject.GetComponent().PointStartID = spawnPoint.PointId; + bots.Add(gameobject); } } + public void ReSpawn(ICharacter character, ref Vector3 pos, ref int startPointId) + { + character.ResetCharacter(); + var team = character.GetCharacter.Team; + NavPoint navPoint; + if (team == Team.Attackers) + navPoint = spawnPointsForAttackersTeam[Random.Range(0, spawnPointsForAttackersTeam.Count)]; + else + navPoint = spawnPointsForDefendersTeam[Random.Range(0, spawnPointsForDefendersTeam.Count)]; + + pos = navPoint.Position; + startPointId = navPoint.PointId; + } + private void ResetCharacters() { - foreach (var bot in Bots) + foreach (var bot in bots) { var npc = bot.GetComponent(); npc.ResetCharacter(); if (npc.GetCharacter.Team == Team.Attackers) - bot.transform.position = spawnPointsForAttackersTeam[Random.Range(0, spawnPointsForAttackersTeam.Count)].position; + bot.transform.position = spawnPointsForAttackersTeam[Random.Range(0, spawnPointsForAttackersTeam.Count)].Position; else - bot.transform.position = spawnPointsForDefendersTeam[Random.Range(0, spawnPointsForDefendersTeam.Count)].position; + bot.transform.position = spawnPointsForDefendersTeam[Random.Range(0, spawnPointsForDefendersTeam.Count)].Position; } Player player; if (TryGetComponent(out player)) { player.ResetCharacter(); if (player.GetCharacter.Team == Team.Attackers) - Player.transform.position = spawnPointsForAttackersTeam[Random.Range(0, spawnPointsForAttackersTeam.Count)].position; + this.player.transform.position = spawnPointsForAttackersTeam[Random.Range(0, spawnPointsForAttackersTeam.Count)].Position; else - Player.transform.position = spawnPointsForDefendersTeam[Random.Range(0, spawnPointsForDefendersTeam.Count)].position; + this.player.transform.position = spawnPointsForDefendersTeam[Random.Range(0, spawnPointsForDefendersTeam.Count)].Position; } } } \ No newline at end of file diff --git a/Assets/Scripts/Bots/TeamEnum.cs b/Assets/Scripts/Bots/TeamEnum.cs index 68279f5..a2c8a95 100755 --- a/Assets/Scripts/Bots/TeamEnum.cs +++ b/Assets/Scripts/Bots/TeamEnum.cs @@ -2,4 +2,15 @@ { Defenders, Attackers, +} + +public static class TeamExtension +{ + public static Team GetOppositeTeam(this Team team) + { + if (team == Team.Attackers) + return Team.Defenders; + else + return Team.Attackers; + } } \ No newline at end of file diff --git a/Assets/Scripts/Character/Character.cs b/Assets/Scripts/Character/Character.cs index 50e7313..3713998 100644 --- a/Assets/Scripts/Character/Character.cs +++ b/Assets/Scripts/Character/Character.cs @@ -7,12 +7,6 @@ public class Character public Character() { - Debug.Log("init"); Condition = new CharacterCondition(); } -} - -public interface ICharacter -{ - Character GetCharacter { get; } } \ No newline at end of file diff --git a/Assets/Scripts/Character/CharacterCondition.cs b/Assets/Scripts/Character/CharacterCondition.cs index b674222..de60740 100755 --- a/Assets/Scripts/Character/CharacterCondition.cs +++ b/Assets/Scripts/Character/CharacterCondition.cs @@ -9,17 +9,17 @@ public class CharacterCondition public event Action OnChangeAmmunitionEvent; private int health; - public int HealthPoints - { - get - { - return health; - } + public int HealthPoints + { + get + { + return health; + } private set { health = value; OnChangeHealthEvent?.Invoke(value); - } + } } public int GetHealthPointsInQuantile() @@ -30,7 +30,7 @@ public class CharacterCondition return 1; else if (health < 75) return 2; - else return 3; + else return 3; } private int armour; public int ArmourPoints @@ -45,6 +45,17 @@ public class CharacterCondition OnChangeArmourEvent?.Invoke(value); } } + public int GetArmourPointsInQuantile() + { + if (armour < 25) + return 0; + else if (armour < 50) + return 1; + else if (armour < 75) + return 2; + else return 3; + } + private int ammo; public int Ammunition { @@ -60,6 +71,11 @@ public class CharacterCondition } public CharacterCondition() + { + this.Reset(); + } + + public void Reset() { var settings = SettingsReader.Instance.GetSettings; ammo = settings.MaxAmmo; diff --git a/Assets/Scripts/Sensors.meta b/Assets/Scripts/Character/Interfaces.meta old mode 100755 new mode 100644 similarity index 77% rename from Assets/Scripts/Sensors.meta rename to Assets/Scripts/Character/Interfaces.meta index a808c2b..cb4a3aa --- a/Assets/Scripts/Sensors.meta +++ b/Assets/Scripts/Character/Interfaces.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: 5e73ba257bc6b684c86edf9ecfd475ef +guid: f23b6db3be1e4cd469fd18dfe3e39764 folderAsset: yes DefaultImporter: externalObjects: {} diff --git a/Assets/Scripts/Character/Interfaces/ICharacter.cs b/Assets/Scripts/Character/Interfaces/ICharacter.cs new file mode 100644 index 0000000..aef14a7 --- /dev/null +++ b/Assets/Scripts/Character/Interfaces/ICharacter.cs @@ -0,0 +1,5 @@ +public interface ICharacter +{ + Character GetCharacter { get; } + void ResetCharacter(); +} \ No newline at end of file diff --git a/Assets/Scripts/Sensors/SensorType.cs.meta b/Assets/Scripts/Character/Interfaces/ICharacter.cs.meta old mode 100755 new mode 100644 similarity index 83% rename from Assets/Scripts/Sensors/SensorType.cs.meta rename to Assets/Scripts/Character/Interfaces/ICharacter.cs.meta index 26f4b58..b53c34e --- a/Assets/Scripts/Sensors/SensorType.cs.meta +++ b/Assets/Scripts/Character/Interfaces/ICharacter.cs.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: 8f76201fe6436164789d10350a0fd6e2 +guid: b6dfb78244ae35c4db1326d5f5b73375 MonoImporter: externalObjects: {} serializedVersion: 2 diff --git a/Assets/Scripts/Character/Interfaces/INpcBaseState.cs b/Assets/Scripts/Character/Interfaces/INpcBaseState.cs new file mode 100644 index 0000000..bf17ad2 --- /dev/null +++ b/Assets/Scripts/Character/Interfaces/INpcBaseState.cs @@ -0,0 +1,17 @@ +using UnityEngine; + +public interface INpcBaseState +{ + NpcEnumState State { get; } + bool InCover { get; } + bool IsRunning { get; } + bool InDirectPoint { get; } + float HitChance { get; } + float DoDamageChance { get; } +} + +public interface INpcBaseBodyState +{ + NpcBodyState State { get; } + Vector3 GetPointToHit(GameObject go); +} \ No newline at end of file diff --git a/Assets/Scripts/Sensors/Sensors.cs.meta b/Assets/Scripts/Character/Interfaces/INpcBaseState.cs.meta similarity index 83% rename from Assets/Scripts/Sensors/Sensors.cs.meta rename to Assets/Scripts/Character/Interfaces/INpcBaseState.cs.meta index 1109bef..f0585e4 100644 --- a/Assets/Scripts/Sensors/Sensors.cs.meta +++ b/Assets/Scripts/Character/Interfaces/INpcBaseState.cs.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: 4599c57bc5b1c3945847dead0f9f0ba4 +guid: 58b7e1962495ada4c8e6ee6219c99a20 MonoImporter: externalObjects: {} serializedVersion: 2 diff --git a/Assets/Scripts/Character/MovementController.cs b/Assets/Scripts/Character/MovementController.cs index 334b7d2..2319c08 100644 --- a/Assets/Scripts/Character/MovementController.cs +++ b/Assets/Scripts/Character/MovementController.cs @@ -1,24 +1,32 @@ -using System.Linq; -using System.Collections.Generic; +using System.Collections.Generic; +using System.Linq; using UnityEngine; using UnityEngine.AI; -using System.Threading.Tasks; [RequireComponent(typeof(NavMeshAgent))] public class MovementController : MonoBehaviour { - public NavPoint CurrentNavPoint { get; set; } + public int PointStartID { get; set; } + public int PointEndID { get; private set; } public float FlagDistance { get; private set; } - private GameObject flag; - private const float updateFlagPositionDelay = 5; - [SerializeField] private NavMeshAgent navMeshAgent; + private const float updateFlagPositionDelay = 5; + private const float updateReachedDestinationDelay = 5; - private void Start() + [SerializeField] private NavMeshAgent navMeshAgent; + [SerializeField] private GameObject flag; + public float DistanceToGo { get; private set; } + public float RemainingDistance => navMeshAgent.remainingDistance; + private Dictionary idNavPointDict; + + + private void Awake() { navMeshAgent.speed = SettingsReader.Instance.GetSettings.MovementSpeed; + idNavPointDict = MapManager.IDToNavPoint; InvokeRepeating(nameof(UpdateFlagPosition), 0, updateFlagPositionDelay); + InvokeRepeating(nameof(ReachedDestination), 0, updateReachedDestinationDelay); } - + private void OnDestroy() { CancelInvoke(nameof(UpdateFlagPosition)); @@ -30,18 +38,46 @@ public class MovementController : MonoBehaviour } public void MoveToRandomPoint() - { - Debug.Log(MapManager.navPoints == null); - goToNextNavPoint(MapManager.navPoints[Random.Range(0, MapManager.navPoints.Count)]); + { + Debug.Log(MapManager.NavPoints == null); + GoToNextNavPoint(MapManager.NavPoints[Random.Range(0, MapManager.NavPoints.Count)]); } - public List getPointsCandidate() + public List GetPointsCandidate() { - return MapManager.navPoints - .Where(point => (CurrentNavPoint.position - point.position).magnitude < SettingsReader.Instance.GetSettings.MovementSpeed) + return MapManager.NavPoints + .Where(point => + (idNavPointDict[PointStartID].Position - point.Position).magnitude < SettingsReader.Instance.GetSettings.MovementDistance) .ToList(); } - public void goToNextNavPoint(NavPoint destination) => - navMeshAgent.SetDestination(destination.position); + public void GoToNextNavPoint(NavPoint destination) + { + if (navMeshAgent.isStopped == true) navMeshAgent.isStopped = false; + PointStartID = PointEndID; + PointEndID = destination.PointId; + navMeshAgent.SetDestination(destination.Position); + DistanceToGo = navMeshAgent.remainingDistance; + } + + public void ReturnToStartPoint() + { + if (navMeshAgent.isStopped == true) navMeshAgent.isStopped = false; + navMeshAgent.SetDestination(idNavPointDict[PointStartID].Position); + PointEndID = PointStartID; + PointStartID = -1; + } + + public void StopOnPath() + { + navMeshAgent.isStopped = true; + PointStartID = -1; + PointEndID = -1; + } + + public void ReachedDestination() + { + if ((navMeshAgent.isStopped == false) && (navMeshAgent.velocity.magnitude < 0.1)) + PointStartID = PointEndID; + } } diff --git a/Assets/Scripts/Character/NPC.cs b/Assets/Scripts/Character/NPC.cs index c738d0b..d47e377 100644 --- a/Assets/Scripts/Character/NPC.cs +++ b/Assets/Scripts/Character/NPC.cs @@ -1,108 +1,170 @@ using System; -using UnityEngine; +using System.Collections.Generic; using Unity.MLAgents; -using Unity.MLAgents.Sensors; using Unity.MLAgents.Actuators; +using Unity.MLAgents.Sensors; +using UnityEngine; -[RequireComponent(typeof(MovementController))] +[RequireComponent(typeof(MovementController),typeof(BufferSensor))] public class NPC : Agent, ICharacter { [HideInInspector] - public Character AgentCharacter; + private Character AgentCharacter; public CharacterCondition Condition; - private FlagZone flagZone; + private FlagZone flagZone = null; - public NPC_BaseState NPC_State { get; private set; } + public INpcBaseState NpcState { get; private set; } + public INpcBaseBodyState NpcBodyState { get; private set; } public Character GetCharacter => AgentCharacter; - private NPC_DirectPointState DirectState; - private NPC_InCoverState CoverState; - private NPC_RunningState RunningState; + private NpcDirectPointState DirectState; + private NpcInCoverState CoverState; + private NpcRunningState RunningState; + + private NpcStandingState StandingState; + private NpcCrouchingState CrouchingState; private MovementController moveController; private BufferSensorComponent bufferSensor; + private Dictionary navPointIdDict; + + #region UnityEvents and ML private void Awake() { - DirectState = new NPC_DirectPointState(); - CoverState = new NPC_InCoverState(); - RunningState = new NPC_RunningState(); - NPC_State = DirectState; + DirectState = new NpcDirectPointState(); + CoverState = new NpcInCoverState(); + RunningState = new NpcRunningState(); + NpcState = DirectState; + + CrouchingState = new NpcCrouchingState(); + StandingState = new NpcStandingState(); + NpcBodyState = StandingState; AgentCharacter = new Character(); Condition = AgentCharacter.Condition; moveController = gameObject.GetComponent(); bufferSensor = gameObject.GetComponent(); - } - - public void ResetCharacter() + flagZone = GameObject.FindObjectOfType(); + if (flagZone == null) + Debug.LogError("Flag Is Not Setted"); + + navPointIdDict = MapManager.IDToNavPoint; + if (navPointIdDict is null) + Debug.LogError("Cant Find Nav Point Dictionary"); + } + + private void OnDestroy() { - Condition = new CharacterCondition(); - EndEpisode(); + Debug.LogWarning("Pooled object was destroyed"); } public override void OnEpisodeBegin() - { - NPC_State = DirectState; + { + NpcState = DirectState; flagZone = GameObject.FindObjectOfType(); } public override void CollectObservations(VectorSensor sensor) { - var candidates = moveController.getPointsCandidate(); + var candidates = moveController.GetPointsCandidate(); - sensor.AddObservation(Condition.HealthPoints); - sensor.AddObservation(Condition.ArmourPoints); - sensor.AddObservation(Condition.Ammunition); - sensor.AddObservation((int)NPC_State.State); - sensor.AddObservation((!flagZone.isNotOccup).ToInt()); + //common sensors + sensor.AddObservation(GameManager.IsHaveSeenByEnemy(AgentCharacter.Team.GetOppositeTeam(), + NpcBodyState.GetPointToHit(gameObject)).ToInt()); sensor.AddObservation(AgentCharacter.LastTimeHit); + sensor.AddObservation((!flagZone.IsNotOccup).ToInt()); sensor.AddObservation(Condition.GetHealthPointsInQuantile()); + sensor.AddObservation(Condition.GetArmourPointsInQuantile()); sensor.AddObservation(candidates.Count); - sensor.AddObservation(GameManager.IsEnemyNearby(gameObject.transform.position, AgentCharacter.Team)); - + sensor.AddObservation(moveController.PointStartID); + sensor.AddObservation(moveController.PointEndID); + //state sensors + sensor.AddObservation((int)NpcState.State); + sensor.AddObservation((int)NpcBodyState.State); + sensor.AddObservation(GameManager.IsEnemyNearby(gameObject.transform.position, AgentCharacter.Team)); + sensor.AddObservation(navPointIdDict[moveController.PointStartID].DeathAttr); + sensor.AddObservation(navPointIdDict[moveController.PointEndID].DeathAttr); + sensor.AddObservation(moveController.FlagDistance); + + //point sensors foreach (var point in candidates) { - Debug.Log((float)moveController.CurrentNavPoint.PointId); - bufferSensor.AppendObservation(new float[] { - //1 position in navpointId - (float)moveController.CurrentNavPoint.PointId, - //2 distance to flag - moveController.FlagDistance, - //3 death count in point - moveController.CurrentNavPoint.DeathAttr, + point.DeathAttr, + (int)point.navType, //4 flagEnemyDistance GameManager.IsCloserToFlagFromNextNavPoint(point, transform.position).ToInt(), //5 EnemyVsNavPointDistance - GameManager.IsCloserToEnemyThanToNextNavPoint(point,transform.position, AgentCharacter.Team).ToInt() - }); - - } - } - - public override void Heuristic(in ActionBuffers actionsOut) - { - var discreteActionsOut = actionsOut.DiscreteActions; - if (Input.GetKeyDown(KeyCode.W)) - { - discreteActionsOut[0] = 1; + GameManager.IsCloserToEnemyThanToNextNavPoint(point,transform.position, AgentCharacter.Team.GetOppositeTeam()).ToInt(), + //6 Have been seen by enemy in this point + GameManager.IsHaveSeenByEnemy(AgentCharacter.Team.GetOppositeTeam(), + point.Position).ToInt() + }); } } public override void OnActionReceived(ActionBuffers actions) { - if (actions.DiscreteActions[0] == 1) + var result = actions.DiscreteActions; + if (result[0] == 0) { - moveController.MoveToRandomPoint(); - NPC_State = RunningState; + if (navPointIdDict[moveController.PointStartID].navType != NavPointType.Cover) + return; + NpcState = CoverState; + + switch (result[1]) + { + case 0: Peek(); break; + case 1: Cover(); break; + case 3: Peek(); moveController.GoToNextNavPoint(navPointIdDict[result[2]]); break; + case 4: NpcState = DirectState; break; + default: throw new ArgumentException("Undefined Action recieved"); + } + } + if (result[0] == 1) + { + if (navPointIdDict[moveController.PointStartID].navType != NavPointType.Direction) + return; + switch (result[1]) + { + case 0: moveController.GoToNextNavPoint(navPointIdDict[result[2]]); + NpcState = RunningState; break; + case 1: NpcState = DirectState; break; + default: throw new ArgumentException("Undefined Action recieved"); + } + } + if (result[0] == 2) + { + if (moveController.PointStartID == moveController.PointEndID && moveController.PointEndID != -1) + return; + switch (result[1]) + { + case 0: moveController.StopOnPath(); NpcState = DirectState; break; + case 1: moveController.ReturnToStartPoint(); NpcState = RunningState; break; + default: throw new ArgumentException("Undefined Action recieved"); + } } } + #endregion - public event Action OnKilledEvent; + public event Action OnChangePosition; + private void Peek() + { + OnChangePosition?.Invoke(global::NpcBodyState.Standing); + NpcBodyState = StandingState; + } + + private void Cover() + { + OnChangePosition?.Invoke(global::NpcBodyState.Crouching); + NpcBodyState = CrouchingState; + } + + public event Action OnDamageRecieved; public void GetDamage(float damage) { AgentCharacter.LastTimeHit = TimeManager.Instance.CurrentTime; @@ -111,13 +173,17 @@ public class NPC : Agent, ICharacter if (Condition.HealthPoints < 0) { - OnKilledEvent?.Invoke(this); - moveController.CurrentNavPoint.DeathAttr += 1; + MapManager.AddDeathAttributeToPoints(moveController.PointStartID, moveController.PointEndID, + moveController.DistanceToGo, moveController.RemainingDistance); + var pos = gameObject.transform.position; + var id = moveController.PointStartID; + CharacterFactory.Instance.ReSpawn(this, ref pos, ref id); } } - private void OnDestroy() + public void ResetCharacter() { - Debug.LogWarning("Pooled object was destroyed"); + Condition.Reset(); + EndEpisode(); } } diff --git a/Assets/Scripts/Character/NPC_State.cs b/Assets/Scripts/Character/NPC_State.cs deleted file mode 100644 index cc2802c..0000000 --- a/Assets/Scripts/Character/NPC_State.cs +++ /dev/null @@ -1,46 +0,0 @@ -public enum NPC_EnumState -{ - InCover, - InDirectPoint, - InRunning, -} - -public interface NPC_BaseState -{ - NPC_EnumState State { get; } - bool InCover { get; } - bool IsRunning { get; } - bool InDirectPoint { get; } - float HitChance { get; } - float DoDamageChance { get; } -} - -public class NPC_DirectPointState : NPC_BaseState -{ - public bool InCover => false; - public bool IsRunning => false; - public bool InDirectPoint => false; - public float HitChance => SettingsReader.Instance.GetSettings.GetHitChanceInDirectPoint; - public float DoDamageChance => SettingsReader.Instance.GetSettings.DoDamageChanceInDirectPoint; - public NPC_EnumState State => NPC_EnumState.InDirectPoint; -} - -public class NPC_RunningState : NPC_BaseState -{ - public bool InCover => false; - public bool IsRunning => true; - public bool InDirectPoint => false; - public float HitChance => SettingsReader.Instance.GetSettings.GetHitChanceInRunning; - public float DoDamageChance => SettingsReader.Instance.GetSettings.DoDamageChanceInRunning; - public NPC_EnumState State => NPC_EnumState.InRunning; -} - -public class NPC_InCoverState : NPC_BaseState -{ - public bool InCover => true; - public bool IsRunning => false; - public bool InDirectPoint => false; - public float HitChance => SettingsReader.Instance.GetSettings.GetHitChanceInCover; - public float DoDamageChance => SettingsReader.Instance.GetSettings.DoDamageChanceInCover; - public NPC_EnumState State => NPC_EnumState.InCover; -} diff --git a/Assets/Scripts/Character/NpcState.cs b/Assets/Scripts/Character/NpcState.cs new file mode 100644 index 0000000..51feda6 --- /dev/null +++ b/Assets/Scripts/Character/NpcState.cs @@ -0,0 +1,68 @@ +using UnityEngine; + +public enum NpcEnumState +{ + InCover, + InDirectPoint, + InRunning, +} + +public enum NpcBodyState +{ + Crouching, + Standing, +} + +public class NpcCrouchingState : INpcBaseBodyState +{ + public NpcBodyState State => NpcBodyState.Crouching; + + public Vector3 GetPointToHit(GameObject go) + { + MeshRenderer meshRenderer; + go.TryGetComponent(out meshRenderer); + return meshRenderer.bounds.center; + } +} + +public class NpcStandingState : INpcBaseBodyState +{ + public NpcBodyState State => NpcBodyState.Standing; + + public Vector3 GetPointToHit(GameObject go) + { + MeshRenderer meshRenderer; + go.TryGetComponent(out meshRenderer); + return meshRenderer.bounds.center; + } +} + +public class NpcDirectPointState : INpcBaseState +{ + public bool InCover => false; + public bool IsRunning => false; + public bool InDirectPoint => false; + public float HitChance => SettingsReader.Instance.GetSettings.GetHitChanceInDirectPoint; + public float DoDamageChance => SettingsReader.Instance.GetSettings.DoDamageChanceInDirectPoint; + public NpcEnumState State => NpcEnumState.InDirectPoint; +} + +public class NpcRunningState : INpcBaseState +{ + public bool InCover => false; + public bool IsRunning => true; + public bool InDirectPoint => false; + public float HitChance => SettingsReader.Instance.GetSettings.GetHitChanceInRunning; + public float DoDamageChance => SettingsReader.Instance.GetSettings.DoDamageChanceInRunning; + public NpcEnumState State => NpcEnumState.InRunning; +} + +public class NpcInCoverState : INpcBaseState +{ + public bool InCover => true; + public bool IsRunning => false; + public bool InDirectPoint => false; + public float HitChance => SettingsReader.Instance.GetSettings.GetHitChanceInCover; + public float DoDamageChance => SettingsReader.Instance.GetSettings.DoDamageChanceInCover; + public NpcEnumState State => NpcEnumState.InCover; +} diff --git a/Assets/Scripts/Character/NPC_State.cs.meta b/Assets/Scripts/Character/NpcState.cs.meta similarity index 100% rename from Assets/Scripts/Character/NPC_State.cs.meta rename to Assets/Scripts/Character/NpcState.cs.meta diff --git a/Assets/Scripts/Character/Player.cs b/Assets/Scripts/Character/Player.cs index e593f51..593a326 100644 --- a/Assets/Scripts/Character/Player.cs +++ b/Assets/Scripts/Character/Player.cs @@ -15,9 +15,9 @@ public class Player : MonoBehaviour, ICharacter Condition = PlayerCharacter.Condition; } - public void ResetCharacter() + private void OnDestroy() { - Condition = new CharacterCondition(); + Debug.LogWarning("Pooled object was destroyed"); } public event Action OnKilledEvent; @@ -27,12 +27,12 @@ public class Player : MonoBehaviour, ICharacter Condition.GiveHealth(-Mathf.RoundToInt(damage * (1 - Condition.ArmourPoints * 0.5f))); Condition.GiveArmour(-Mathf.RoundToInt(Mathf.Sqrt(damage) * 5)); - if (Condition.HealthPoints < 0) - OnKilledEvent?.Invoke(this); + if (Condition.HealthPoints < 0) + OnKilledEvent?.Invoke(this); } - private void OnDestroy() + public void ResetCharacter() { - Debug.LogWarning("Pooled object was destroyed"); + Condition = new CharacterCondition(); } } diff --git a/Assets/Scripts/Character/scr_CharacterController.cs b/Assets/Scripts/Character/scr_CharacterController.cs index 2e54f22..e4f847d 100755 --- a/Assets/Scripts/Character/scr_CharacterController.cs +++ b/Assets/Scripts/Character/scr_CharacterController.cs @@ -1,10 +1,6 @@ -using System; -using System.Collections; -using System.Collections.Generic; -using Unity.Barracuda; -using UnityEngine; +using UnityEngine; -using static scr_Models; +using static scr_Models; public class scr_CharacterController : MonoBehaviour { @@ -14,7 +10,7 @@ public class scr_CharacterController : MonoBehaviour private Vector2 input_Movement; [HideInInspector] public Vector2 input_View; - + private Vector3 newCameraRotation; private Vector3 newCharacterRotation; @@ -22,14 +18,14 @@ public class scr_CharacterController : MonoBehaviour public Transform cameraHolder; public Transform feetTransform; - [Header("Settings")] + [Header("Settings")] public PlayerSettingsModel playerSettings; public float ViewClampYMin = -70; public float ViewClampYMax = 80; public LayerMask playerMask; - - [Header("Gravity")] + + [Header("Gravity")] public float gravityAmount; public float gravityMin; private float playerGravity; @@ -37,14 +33,14 @@ public class scr_CharacterController : MonoBehaviour public Vector3 jumpingForce; private Vector3 jumpingForceVelocity; - [Header("Stance")] + [Header("Stance")] public PlayerStance playerStance; public float playerStanceSmoothing; public CharacterStance playerStandStance; public CharacterStance playerCrouchStance; public CharacterStance playerProneStance; private float stanceCheckErrorMargin = 0.05f; - + private float cameraHeight; private float cameraHeightVelocity; @@ -61,13 +57,13 @@ public class scr_CharacterController : MonoBehaviour defaultInput.Character.Movement.performed += e => input_Movement = e.ReadValue(); defaultInput.Character.View.performed += e => input_View = e.ReadValue(); defaultInput.Character.Jump.performed += e => Jump(); - + defaultInput.Character.Crouch.performed += e => Crouch(); defaultInput.Character.Prone.performed += e => Prone(); - + defaultInput.Character.Sprint.performed += e => ToggleSprint(); defaultInput.Character.SprintReleased.performed += e => StopSprint(); - + defaultInput.Enable(); newCameraRotation = cameraHolder.localRotation.eulerAngles; @@ -80,7 +76,7 @@ public class scr_CharacterController : MonoBehaviour { currentWeapon.Initialise(this); } - + } private void Update() @@ -95,10 +91,10 @@ public class scr_CharacterController : MonoBehaviour { newCharacterRotation.y += playerSettings.ViewXSensetivity * (playerSettings.ViewXInverted ? -input_View.x : input_View.x) * Time.deltaTime; transform.localRotation = Quaternion.Euler(newCharacterRotation); - + newCameraRotation.x += playerSettings.ViewYSensetivity * (playerSettings.ViewYInverted ? input_View.y : -input_View.y) * Time.deltaTime; newCameraRotation.x = Mathf.Clamp(newCameraRotation.x, ViewClampYMin, ViewClampYMax); - + cameraHolder.localRotation = Quaternion.Euler(newCameraRotation); } @@ -108,7 +104,7 @@ public class scr_CharacterController : MonoBehaviour { isSprinting = false; } - + var verticalSpeed = playerSettings.WalkingForwardSpeed; var horizontalSpeed = playerSettings.WalkingStrafeSpeed; @@ -117,17 +113,17 @@ public class scr_CharacterController : MonoBehaviour verticalSpeed = playerSettings.RunningForwardSpeed; horizontalSpeed = playerSettings.RunningStrafeSpeed; } - + // Effectors if (!characterController.isGrounded) { playerSettings.SpeedEffector = playerSettings.FallingSpeedEffector; } - else if(playerStance == PlayerStance.Crouch) + else if (playerStance == PlayerStance.Crouch) { playerSettings.SpeedEffector = playerSettings.CrouchSpeedEffector; - } - else if(playerStance == PlayerStance.Prone) + } + else if (playerStance == PlayerStance.Prone) { playerSettings.SpeedEffector = playerSettings.ProneSpeedEffector; } @@ -135,15 +131,15 @@ public class scr_CharacterController : MonoBehaviour { playerSettings.SpeedEffector = 1; } - + verticalSpeed *= playerSettings.SpeedEffector; horizontalSpeed *= playerSettings.SpeedEffector; - + newMovementSpeed = Vector3.SmoothDamp(newMovementSpeed, new Vector3(horizontalSpeed * input_Movement.x * Time.deltaTime, 0, verticalSpeed * input_Movement.y * Time.deltaTime), ref newMovementSpeedVelocity, characterController.isGrounded ? playerSettings.MovementSmoothing : playerSettings.FallingSmoothing); - + var MovementSpeed = transform.TransformDirection(newMovementSpeed); if (playerGravity > gravityMin) @@ -158,7 +154,7 @@ public class scr_CharacterController : MonoBehaviour MovementSpeed.y += playerGravity; MovementSpeed += jumpingForce * Time.deltaTime; - + characterController.Move(MovementSpeed); } @@ -179,7 +175,7 @@ public class scr_CharacterController : MonoBehaviour { stanceHeight = playerProneStance.CameraHeight; } - + cameraHeight = Mathf.SmoothDamp(cameraHolder.localPosition.y, stanceHeight, ref cameraHeightVelocity, playerStanceSmoothing); cameraHolder.localPosition = new Vector3(cameraHolder.localPosition.x, cameraHeight, cameraHolder.localPosition.z); @@ -190,7 +186,7 @@ public class scr_CharacterController : MonoBehaviour { return; } - + if (playerStance == PlayerStance.Crouch) { if (StanceCheck(playerStandStance.StanceCollider.height)) @@ -200,7 +196,7 @@ public class scr_CharacterController : MonoBehaviour playerStance = PlayerStance.Stand; return; } - + // Jump jumpingForce = Vector3.up * playerSettings.JumpingHeight; playerGravity = 0; @@ -233,8 +229,8 @@ public class scr_CharacterController : MonoBehaviour { var start = new Vector3(feetTransform.position.x, feetTransform.position.y + characterController.radius + stanceCheckErrorMargin, feetTransform.position.z); var end = new Vector3(feetTransform.position.x, feetTransform.position.y - characterController.radius - stanceCheckErrorMargin + stanceCheckheight, feetTransform.position.z); - - + + return Physics.CheckCapsule(start, end, characterController.radius, playerMask); } @@ -247,7 +243,7 @@ public class scr_CharacterController : MonoBehaviour } isSprinting = !isSprinting; } - + private void StopSprint() { if (playerSettings.SprintingHold) @@ -255,5 +251,5 @@ public class scr_CharacterController : MonoBehaviour isSprinting = false; } } - + } diff --git a/Assets/Scripts/Character/scr_Models.cs b/Assets/Scripts/Character/scr_Models.cs index 3aca984..c735db5 100755 --- a/Assets/Scripts/Character/scr_Models.cs +++ b/Assets/Scripts/Character/scr_Models.cs @@ -3,73 +3,73 @@ using UnityEngine; public static class scr_Models { - #region Player - - public enum PlayerStance - { - Stand, - Crouch, - Prone - } - - [Serializable] - public class PlayerSettingsModel - { - [Header("View Settings")] - public float ViewXSensetivity; - public float ViewYSensetivity; + #region Player - public bool ViewXInverted; - public bool ViewYInverted; + public enum PlayerStance + { + Stand, + Crouch, + Prone + } - [Header("Movement Settings")] - public bool SprintingHold; - public float MovementSmoothing; - - [Header("Movement - Running")] - public float RunningForwardSpeed; - public float RunningStrafeSpeed; - - [Header("Movement - Walking")] - public float WalkingForwardSpeed; - public float WalkingBackwardSpeed; - public float WalkingStrafeSpeed; + [Serializable] + public class PlayerSettingsModel + { + [Header("View Settings")] + public float ViewXSensetivity; + public float ViewYSensetivity; - [Header("Jumping")] - public float JumpingHeight; - public float JumpingFalloff; - public float FallingSmoothing; + public bool ViewXInverted; + public bool ViewYInverted; - [Header("Speed Effectors")] - public float SpeedEffector = 1; - public float CrouchSpeedEffector; - public float ProneSpeedEffector; - public float FallingSpeedEffector; - } + [Header("Movement Settings")] + public bool SprintingHold; + public float MovementSmoothing; - [Serializable] - public class CharacterStance - { - public float CameraHeight; - public CapsuleCollider StanceCollider; - } + [Header("Movement - Running")] + public float RunningForwardSpeed; + public float RunningStrafeSpeed; - #endregion + [Header("Movement - Walking")] + public float WalkingForwardSpeed; + public float WalkingBackwardSpeed; + public float WalkingStrafeSpeed; - #region - Weapons - + [Header("Jumping")] + public float JumpingHeight; + public float JumpingFalloff; + public float FallingSmoothing; - [Serializable] - public class WeaponSettingsModel - { - [Header("Sway")] - public float SwayAmount; - public bool SwayYInverted; - public bool SwayXInverted; - public float SwaySmoothing; - public float SwayResetSmoothing; - public float SwayClampX; - public float SwayClampY; - } + [Header("Speed Effectors")] + public float SpeedEffector = 1; + public float CrouchSpeedEffector; + public float ProneSpeedEffector; + public float FallingSpeedEffector; + } - #endregion + [Serializable] + public class CharacterStance + { + public float CameraHeight; + public CapsuleCollider StanceCollider; + } + + #endregion + + #region - Weapons - + + [Serializable] + public class WeaponSettingsModel + { + [Header("Sway")] + public float SwayAmount; + public bool SwayYInverted; + public bool SwayXInverted; + public float SwaySmoothing; + public float SwayResetSmoothing; + public float SwayClampX; + public float SwayClampY; + } + + #endregion } diff --git a/Assets/Scripts/Managers/GameManager.cs b/Assets/Scripts/Managers/GameManager.cs index b91908d..7720a90 100755 --- a/Assets/Scripts/Managers/GameManager.cs +++ b/Assets/Scripts/Managers/GameManager.cs @@ -1,76 +1,123 @@ -using Unity.MLAgents; +using System; +using Unity.MLAgents; using UnityEngine; -using System; public class GameManager : MonoBehaviour { private static GameManager instance; - public static GameManager Instance { get { return instance; } } + public static GameManager Instance => instance; - private static SimpleMultiAgentGroup DefendersTeam = new SimpleMultiAgentGroup(); - private static SimpleMultiAgentGroup AttackersTeam = new SimpleMultiAgentGroup(); + private static SimpleMultiAgentGroup defendersTeam = new SimpleMultiAgentGroup(); + private static SimpleMultiAgentGroup attackersTeam = new SimpleMultiAgentGroup(); private void Awake() { - if (Instance == null) + if (instance is null) instance = this; - else if (Instance == this) + else + { Destroy(gameObject); + Debug.LogError("Only 1 Instance"); + } } private void Start() { Academy.Instance.OnEnvironmentReset += ResetScene; - GlobalEventManager.onCaptureFlag += flagCaptured; - GlobalEventManager.onTimeLeft += timeOut; + GlobalEventManager.onCaptureFlag += FlagCaptured; + GlobalEventManager.onTimeLeft += TimeOut; var agents = GameObject.FindObjectsOfType(); foreach (var item in agents) { var agent = item as NPC; if (agent.GetCharacter.Team == Team.Attackers) - AttackersTeam.RegisterAgent(agent); + attackersTeam.RegisterAgent(item); else - DefendersTeam.RegisterAgent(agent); + defendersTeam.RegisterAgent(item); } } - public static bool IsCloserToEnemyThanToNextNavPoint(NavPoint navPoint, Vector3 currentTransform, Team team) + private static SimpleMultiAgentGroup getAgentList(Team team) { - SimpleMultiAgentGroup agentGroup; if (team == Team.Attackers) - agentGroup = AttackersTeam; + return attackersTeam; else - agentGroup = DefendersTeam; + return defendersTeam; + } - var distToNavPoint = (currentTransform - navPoint.position).magnitude; + public static bool IsCloserToEnemyThanToNextNavPoint(NavPoint navPoint, Vector3 currentTransform, Team oppositeTeam) + { + var agentGroup = getAgentList(oppositeTeam); + + var distToNavPoint = (currentTransform - navPoint.Position).magnitude; foreach (var agent in agentGroup.GetRegisteredAgents()) if (distToNavPoint > (currentTransform - agent.transform.position).magnitude) return true; + if ((SettingsReader.Instance.GetSettings.HasHumanAttacker == true && oppositeTeam == Team.Attackers) || + (SettingsReader.Instance.GetSettings.HasHumanDefender == true && oppositeTeam == Team.Defenders)) + { + if (distToNavPoint > (currentTransform - CharacterFactory.Instance.player.transform.position).magnitude) + return true; + } return false; } - public static bool IsEnemyNearby(Vector3 currentTransform, Team team) + public static bool IsEnemyNearby(Vector3 currentTransform, Team oppositeTeam) { - SimpleMultiAgentGroup agentGroup; - if (team == Team.Attackers) - agentGroup = AttackersTeam; - else - agentGroup = DefendersTeam; + var agentGroup = getAgentList(oppositeTeam); foreach (var agent in agentGroup.GetRegisteredAgents()) if ((currentTransform - agent.transform.position).magnitude < SettingsReader.Instance.GetSettings.ViewDistance) return true; + if ((SettingsReader.Instance.GetSettings.HasHumanAttacker == true && oppositeTeam == Team.Attackers) || + (SettingsReader.Instance.GetSettings.HasHumanDefender == true && oppositeTeam == Team.Defenders)) + { + if ((currentTransform - CharacterFactory.Instance.player.transform.position).magnitude < SettingsReader.Instance.GetSettings.ViewDistance) + return true; + } return false; } public static bool IsCloserToFlagFromNextNavPoint(NavPoint navPoint, Vector3 currentTransform) => navPoint.FlagDistance < (currentTransform - GameObject.FindGameObjectWithTag("Flag").transform.position).magnitude; - private void flagCaptured(Team team) + public static bool IsHaveSeenByEnemy(Team oppositeTeam, Vector3 position) { - switch(team) + var agentGroup = getAgentList(oppositeTeam); + RaycastHit rayHit = new RaycastHit(); + foreach (var agent in agentGroup.GetRegisteredAgents() ) + { + var npc = agent as NPC; + if (Physics.Raycast(position, + (npc.NpcBodyState.GetPointToHit(npc.gameObject) - position).normalized, + out rayHit, + SettingsReader.Instance.GetSettings.ViewDistance)) + { + if (rayHit.collider.gameObject.GetComponent() != null) + return true; + } + } + if ((SettingsReader.Instance.GetSettings.HasHumanAttacker == true && oppositeTeam == Team.Attackers) || + (SettingsReader.Instance.GetSettings.HasHumanDefender == true && oppositeTeam == Team.Defenders)) + { + var player = CharacterFactory.Instance.player; + if (Physics.Raycast(position, + (player.GetComponent().bounds.center - position).normalized, + out rayHit, + SettingsReader.Instance.GetSettings.ViewDistance)) + { + if (rayHit.collider.gameObject.GetComponent() != null) + return true; + } + } + return false; + } + + private void FlagCaptured(Team team) + { + switch (team) { case Team.Attackers: Debug.Log("Attackers Win"); @@ -78,21 +125,19 @@ public class GameManager : MonoBehaviour case Team.Defenders: Debug.Log("Defenders Win"); break; - default: - Debug.LogError("Unexpected Team"); - break; } + ResetScene(); } - private void timeOut() + private void TimeOut() { - Debug.Log("Time is out"); + ResetScene(); } private void OnDestroy() { - GlobalEventManager.onCaptureFlag -= flagCaptured; - GlobalEventManager.onTimeLeft -= timeOut; + GlobalEventManager.onCaptureFlag -= FlagCaptured; + GlobalEventManager.onTimeLeft -= TimeOut; } public static event Action OnResetScene; diff --git a/Assets/Scripts/Managers/MapManager.cs b/Assets/Scripts/Managers/MapManager.cs index 8c64208..b28dcba 100755 --- a/Assets/Scripts/Managers/MapManager.cs +++ b/Assets/Scripts/Managers/MapManager.cs @@ -3,17 +3,62 @@ using UnityEngine; public class MapManager : MonoBehaviour { - public static List navPoints { get; private set; } - private void Start() + private static MapManager instance; + public static MapManager Instance => instance; + private static List navPoints = new List(); + private static Dictionary iDToNavPoint = new Dictionary(); + public static List NavPoints { get => navPoints; private set => navPoints = value; } + public static Dictionary IDToNavPoint { get => iDToNavPoint; private set => iDToNavPoint = value; } + + private void Awake() { - var i = 0; - navPoints = new List(); - var navPointsGameObj = GameObject.FindGameObjectsWithTag("Point"); - foreach (var gameobj in navPointsGameObj) + if (instance is null) + instance = this; + else { - var navpoint = gameobj.GetComponent(); - navpoint.PointId = i; i++; - navPoints.Add(navpoint); + Destroy(gameObject); + Debug.LogError("Only 1 Instance"); } } + + private void Start() + { + var navPointSet = GameObject.Find("NavPoint Set"); + var count = navPointSet.transform.childCount; + for (int i=0; i < count; i++) + NavPoints.Add(navPointSet.transform.GetChild(i) + .gameObject.GetComponent()); + + NavPointSetToID(); + } + + private void NavPointSetToID() + { + int i = 0; + foreach (var navPoint in NavPoints) + { + IDToNavPoint.Add(i, navPoint); + navPoint.PointId = i; + i++; + } + } + + public static void AddDeathAttributeToPoints(int startPoint, int endPoint, + float allDistance, float remainingDistance) + { + var startNavPoint = IDToNavPoint[startPoint]; + var endNavPoint = IDToNavPoint[endPoint]; + float coef; + try + { + coef = remainingDistance / allDistance; + } + catch (System.ArithmeticException) + { + Debug.LogError("Path Length is zero"); + return; + } + startNavPoint.DeathAttr += 1 - coef; + endNavPoint.DeathAttr += coef; + } } diff --git a/Assets/Scripts/Managers/TimeManager.cs b/Assets/Scripts/Managers/TimeManager.cs index ea7fec4..4aa32a1 100755 --- a/Assets/Scripts/Managers/TimeManager.cs +++ b/Assets/Scripts/Managers/TimeManager.cs @@ -1,6 +1,4 @@ -using System.Collections; -using System.Collections.Generic; -using UnityEngine; +using UnityEngine; public class TimeManager : MonoBehaviour { @@ -17,12 +15,14 @@ public class TimeManager : MonoBehaviour } else { - Debug.LogError("Only one Instance"); + Debug.LogError("Only 1 Instance"); Destroy(gameObject); } } void Update() { CurrentTime += Time.deltaTime; + if (CurrentTime > SettingsReader.Instance.GetSettings.TimeOut) + GlobalEventManager.SendTimeout(); } } diff --git a/Assets/Scripts/Misc/FlagZone.cs b/Assets/Scripts/Misc/FlagZone.cs index c6328dd..c66d5bb 100755 --- a/Assets/Scripts/Misc/FlagZone.cs +++ b/Assets/Scripts/Misc/FlagZone.cs @@ -1,6 +1,4 @@ -using System.Collections; -using System.Collections.Generic; -using UnityEngine; +using UnityEngine; public class FlagZone : MonoBehaviour { @@ -9,8 +7,8 @@ public class FlagZone : MonoBehaviour public float TimeStayDefenders { get; private set; } private int occupDefenders; private int occupAttackers; - public bool isOccupBoth => (occupDefenders>0) && (occupAttackers>0); - public bool isNotOccup => (occupDefenders == 0) && (occupAttackers == 0); + public bool IsOccupBoth => (occupDefenders > 0) && (occupAttackers > 0); + public bool IsNotOccup => (occupDefenders == 0) && (occupAttackers == 0); private float timeForWin; private void Start() @@ -24,7 +22,7 @@ public class FlagZone : MonoBehaviour } private void OnTriggerEnter(Collider other) { - switch(other.tag) + switch (other.tag) { case "Defender": occupDefenders++; @@ -54,7 +52,7 @@ public class FlagZone : MonoBehaviour } private void Update() { - if (isOccupBoth || isNotOccup) + if (IsOccupBoth || IsNotOccup) { TimeStayAttackers = 0; TimeStayDefenders = 0; @@ -64,7 +62,7 @@ public class FlagZone : MonoBehaviour { TimeStayAttackers += Time.deltaTime; if (TimeStayAttackers > timeForWin) - GlobalEventManager.SendCaptureFlag(Team.Attackers); + GlobalEventManager.SendCaptureFlag(Team.Attackers); } else { diff --git a/Assets/Scripts/Misc/NavPoint.cs b/Assets/Scripts/Misc/NavPoint.cs index 2b43e84..a70fd61 100755 --- a/Assets/Scripts/Misc/NavPoint.cs +++ b/Assets/Scripts/Misc/NavPoint.cs @@ -1,20 +1,28 @@ -using System.Collections; -using System.Collections.Generic; +using System.Collections.Generic; +using System; using UnityEngine; +public enum NavPointType +{ + Cover, + Direction, +} + + public class NavPoint : MonoBehaviour { - public Vector3 position => gameObject.transform.position; + public Vector3 Position => gameObject.transform.position; public float FlagDistance { get; private set; } + public NavPointType navType = NavPointType.Direction; + [HideInInspector] - public int? PointId; + public int PointId = 0; public float DeathAttr = 0; public List EnemiesSeen = new List(); - //Here other attributes; private void Start() { - FlagDistance = (GameObject.FindGameObjectWithTag("Flag").transform.position - position).magnitude; + FlagDistance = (GameObject.FindGameObjectWithTag("Flag").transform.position - Position).magnitude; } } diff --git a/Assets/Scripts/Misc/Settings.cs b/Assets/Scripts/Misc/Settings.cs index b1b7ea8..4e333fc 100755 --- a/Assets/Scripts/Misc/Settings.cs +++ b/Assets/Scripts/Misc/Settings.cs @@ -1,6 +1,6 @@ using UnityEngine; -[CreateAssetMenu(fileName ="Game Settings", menuName = "Game/Settings", order = 51)] +[CreateAssetMenu(fileName = "Game Settings", menuName = "Game/Settings", order = 51)] public class Settings : ScriptableObject { public bool IsTesting; @@ -36,4 +36,6 @@ public class Settings : ScriptableObject public float DoDamageChanceInDirectPoint; public float DoDamageChanceInRunning; public float DoDamageChanceInCover; + + public float CrouchingCoefficient; } diff --git a/Assets/Scripts/Misc/SettingsReader.cs b/Assets/Scripts/Misc/SettingsReader.cs index ab60b87..9e709ad 100755 --- a/Assets/Scripts/Misc/SettingsReader.cs +++ b/Assets/Scripts/Misc/SettingsReader.cs @@ -1,17 +1,21 @@ -using System.Collections; -using System.Collections.Generic; -using UnityEngine; +using UnityEngine; public class SettingsReader : MonoBehaviour { private static SettingsReader instance; - public static SettingsReader Instance { get { return instance; } } + public static SettingsReader Instance => instance; private void Awake() { - instance = this; + if (instance is null) + instance = this; + else + { + Destroy(gameObject); + Debug.LogError("Only 1 Instance"); + } } [SerializeField] private Settings gameSettings; - public Settings GetSettings { get { return gameSettings; } } + public Settings GetSettings => gameSettings; } diff --git a/Assets/Scripts/Misc/Statistics.cs b/Assets/Scripts/Misc/Statistics.cs deleted file mode 100755 index 98be783..0000000 --- a/Assets/Scripts/Misc/Statistics.cs +++ /dev/null @@ -1,9 +0,0 @@ -using UnityEngine; - -public class Statistics : MonoBehaviour -{ - private void Start() - { - - } -} diff --git a/Assets/Scripts/Pickups/AmmoPickUp.cs b/Assets/Scripts/Pickups/AmmoPickUp.cs index 8439c21..e7af1ce 100755 --- a/Assets/Scripts/Pickups/AmmoPickUp.cs +++ b/Assets/Scripts/Pickups/AmmoPickUp.cs @@ -1,5 +1,4 @@ -using System; -using UnityEngine; +using UnityEngine; [RequireComponent(typeof(BoxCollider))] public class AmmoPickUp : MonoBehaviour, IPickable @@ -11,6 +10,11 @@ public class AmmoPickUp : MonoBehaviour, IPickable PickObject(other.gameObject); } + private void OnDestroy() + { + Debug.LogWarning("Pooled object was destroyed"); + } + public void PickObject(GameObject obj) { obj.GetComponent()?.GetCharacter.Condition.TakeAmmo(SettingsReader.Instance.GetSettings.AmmunitionPickupAmount); diff --git a/Assets/Scripts/Pickups/ArmourPickUp.cs b/Assets/Scripts/Pickups/ArmourPickUp.cs index c68e491..b5303f6 100755 --- a/Assets/Scripts/Pickups/ArmourPickUp.cs +++ b/Assets/Scripts/Pickups/ArmourPickUp.cs @@ -1,5 +1,4 @@ -using System; -using UnityEngine; +using UnityEngine; [RequireComponent(typeof(BoxCollider))] public class ArmourPickUp : MonoBehaviour, IPickable @@ -11,6 +10,11 @@ public class ArmourPickUp : MonoBehaviour, IPickable PickObject(other.gameObject); } + private void OnDestroy() + { + Debug.LogWarning("Pooled object was destroyed"); + } + public void PickObject(GameObject obj) { obj.GetComponent()?.GetCharacter.Condition.GiveArmour(SettingsReader.Instance.GetSettings.ArmourPickupAmount); diff --git a/Assets/Scripts/Pickups/HealthPickUp.cs b/Assets/Scripts/Pickups/HealthPickUp.cs index 36f81ba..ba8d136 100755 --- a/Assets/Scripts/Pickups/HealthPickUp.cs +++ b/Assets/Scripts/Pickups/HealthPickUp.cs @@ -1,5 +1,4 @@ -using System; -using UnityEngine; +using UnityEngine; [RequireComponent(typeof(BoxCollider))] public class HealthPickUp : MonoBehaviour, IPickable @@ -11,6 +10,11 @@ public class HealthPickUp : MonoBehaviour, IPickable PickObject(other.gameObject); } + private void OnDestroy() + { + Debug.LogWarning("Pooled object was destroyed"); + } + public void PickObject(GameObject obj) { obj.GetComponent()?.GetCharacter.Condition.GiveHealth(SettingsReader.Instance.GetSettings.HealthPickupAmount); diff --git a/Assets/Scripts/Pickups/IPickable.cs b/Assets/Scripts/Pickups/IPickable.cs index fb218d0..37cced3 100755 --- a/Assets/Scripts/Pickups/IPickable.cs +++ b/Assets/Scripts/Pickups/IPickable.cs @@ -1,7 +1,6 @@ -using System; -using UnityEngine; +using UnityEngine; public interface IPickable { - PickUpType type { get; } + PickUpType type { get; } void PickObject(GameObject obj); } \ No newline at end of file diff --git a/Assets/Scripts/Pickups/PickUpSpawner.cs b/Assets/Scripts/Pickups/PickUpSpawner.cs index 7a29705..f9c320b 100755 --- a/Assets/Scripts/Pickups/PickUpSpawner.cs +++ b/Assets/Scripts/Pickups/PickUpSpawner.cs @@ -44,25 +44,25 @@ public class PickUpSpawner : MonoBehaviour private IEnumerator SpawnNewPickUps() { - while(true) + while (true) { GameObject item; - if(IsDisableCheck(out item)) + if (IsDisableCheck(out item)) { yield return new WaitForSeconds(3); if (item != null) { - item.transform.position = spawnPoints[Random.Range(0, spawnPoints.Count)].position; + item.transform.position = spawnPoints[Random.Range(0, spawnPoints.Count)].Position; item.SetActive(true); } } - yield return new WaitForSeconds(2); + yield return new WaitForSeconds(2); } } private bool IsDisableCheck(out GameObject gameobj) { - foreach(var pick in pickups) + foreach (var pick in pickups) { if (!pick.activeInHierarchy) { diff --git a/Assets/Scripts/Sensors/SensorType.cs b/Assets/Scripts/Sensors/SensorType.cs deleted file mode 100755 index 8d85790..0000000 --- a/Assets/Scripts/Sensors/SensorType.cs +++ /dev/null @@ -1,6 +0,0 @@ -public enum SensorType -{ - Visual, - Sound, - Other -} diff --git a/Assets/Scripts/Sensors/Sensors.cs b/Assets/Scripts/Sensors/Sensors.cs deleted file mode 100644 index b38d85f..0000000 --- a/Assets/Scripts/Sensors/Sensors.cs +++ /dev/null @@ -1,4 +0,0 @@ -using System.Collections.Generic; -using Unity.MLAgents.Sensors; - - diff --git a/Assets/Scripts/Statistics.meta b/Assets/Scripts/Statistics.meta new file mode 100644 index 0000000..4e72120 --- /dev/null +++ b/Assets/Scripts/Statistics.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: 3a9f7f0a9faf11f49a433480722bffc5 +folderAsset: yes +DefaultImporter: + externalObjects: {} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Scripts/Statistics/Logger.cs b/Assets/Scripts/Statistics/Logger.cs new file mode 100644 index 0000000..e293d25 --- /dev/null +++ b/Assets/Scripts/Statistics/Logger.cs @@ -0,0 +1,19 @@ +using System.IO; +using UnityEngine; + +public class Logger +{ + private const string directory = "/Logs/"; + private const string baseName = "Log#"; + + public static void SaveLog(T objToSerialize) + { + string dir = Application.persistentDataPath + directory; + if (!Directory.Exists(dir)) + Directory.CreateDirectory(dir); + + var logName = baseName + (Directory.GetFiles(dir).Length + 1).ToString(); + string json = JsonUtility.ToJson(objToSerialize); + File.WriteAllText(dir + logName, json); + } +} \ No newline at end of file diff --git a/Assets/Scripts/Statistics/Logger.cs.meta b/Assets/Scripts/Statistics/Logger.cs.meta new file mode 100644 index 0000000..e455173 --- /dev/null +++ b/Assets/Scripts/Statistics/Logger.cs.meta @@ -0,0 +1,11 @@ +fileFormatVersion: 2 +guid: b3a1cec894fa98b4bbe20470f1e316c4 +MonoImporter: + externalObjects: {} + serializedVersion: 2 + defaultReferences: [] + executionOrder: 0 + icon: {instanceID: 0} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Scripts/Statistics/StatisticManager.cs b/Assets/Scripts/Statistics/StatisticManager.cs new file mode 100644 index 0000000..399e7fb --- /dev/null +++ b/Assets/Scripts/Statistics/StatisticManager.cs @@ -0,0 +1,51 @@ +using UnityEngine; + +internal class Log +{ + public int damageTakenByDefs = 0; + public int damageTakenByAtc = 0; + + public int AtcWin = 0; + public int DefWin = 0; + + public int TimeOuts = 0; +} + +public class StatisticManager : MonoBehaviour +{ + private Log log = new Log(); + private void Awake() + { + foreach (var npc in GameObject.FindObjectsOfType()) + npc.OnDamageRecieved += RegisterDamage; + + GlobalEventManager.onCaptureFlag += RegisterWin; + GlobalEventManager.onTimeLeft += RegisterTimeOut; + } + + private void RegisterDamage(int damage, Team team) + { + if (team == Team.Attackers) + log.damageTakenByAtc += damage; + else + log.damageTakenByDefs += damage; + } + + private void RegisterWin(Team team) + { + if (team == Team.Attackers) + log.AtcWin += 1; + else + log.DefWin += 1; + } + + private void RegisterTimeOut() + { + log.TimeOuts += 1; + } + + private void OnApplicationQuit() + { + Logger.SaveLog(log); + } +} diff --git a/Assets/Scripts/Misc/Statistics.cs.meta b/Assets/Scripts/Statistics/StatisticManager.cs.meta old mode 100755 new mode 100644 similarity index 91% rename from Assets/Scripts/Misc/Statistics.cs.meta rename to Assets/Scripts/Statistics/StatisticManager.cs.meta index b8be93a..3a27c34 --- a/Assets/Scripts/Misc/Statistics.cs.meta +++ b/Assets/Scripts/Statistics/StatisticManager.cs.meta @@ -4,7 +4,7 @@ MonoImporter: externalObjects: {} serializedVersion: 2 defaultReferences: [] - executionOrder: 0 + executionOrder: 300 icon: {instanceID: 0} userData: assetBundleName: diff --git a/Assets/Scripts/Utils/SerializableDictionary.cs b/Assets/Scripts/Utils/SerializableDictionary.cs index c0877b9..8cb258e 100755 --- a/Assets/Scripts/Utils/SerializableDictionary.cs +++ b/Assets/Scripts/Utils/SerializableDictionary.cs @@ -1,10 +1,10 @@ using System; -using System.Linq; using System.Collections; using System.Collections.Generic; using System.Diagnostics; -using UnityEngine; +using System.Linq; using UnityEditor; +using UnityEngine; using UnityObject = UnityEngine.Object; [Serializable, DebuggerDisplay("Count = {Count}")] diff --git a/Assets/Scripts/Weapons/scr_WeaponController.cs b/Assets/Scripts/Weapons/scr_WeaponController.cs index 2fc59e5..613fc61 100755 --- a/Assets/Scripts/Weapons/scr_WeaponController.cs +++ b/Assets/Scripts/Weapons/scr_WeaponController.cs @@ -1,17 +1,16 @@ -using System; -using UnityEngine; +using UnityEngine; using static scr_Models; public class scr_WeaponController : MonoBehaviour { private scr_CharacterController characterController; - [Header("Settings")] + [Header("Settings")] public WeaponSettingsModel settings; private bool isInitialised; Vector3 newWeaponRotation; Vector3 newWeaponRotationVelocity; - + Vector3 targetWeaponRotation; Vector3 targetWeaponRotationVelocity; @@ -32,17 +31,17 @@ public class scr_WeaponController : MonoBehaviour { return; } - + targetWeaponRotation.y += settings.SwayAmount * (settings.SwayXInverted ? -characterController.input_View.x : characterController.input_View.x) * Time.deltaTime; - targetWeaponRotation.x += settings.SwayAmount * (settings.SwayYInverted ? characterController.input_View.y : -characterController.input_View.y) * Time.deltaTime; + targetWeaponRotation.x += settings.SwayAmount * (settings.SwayYInverted ? characterController.input_View.y : -characterController.input_View.y) * Time.deltaTime; //newWeaponRotation.x = Mathf.Clamp(newWeaponRotation.x, ViewClampYMin, ViewClampYMax); - + targetWeaponRotation.x = Mathf.Clamp(targetWeaponRotation.x, -settings.SwayClampX, settings.SwayClampX); targetWeaponRotation.y = Mathf.Clamp(targetWeaponRotation.y, -settings.SwayClampY, settings.SwayClampY); targetWeaponRotation = Vector3.SmoothDamp(targetWeaponRotation, Vector3.zero, ref targetWeaponRotationVelocity, settings.SwayResetSmoothing); newWeaponRotation = Vector3.SmoothDamp(newWeaponRotation, targetWeaponRotation, ref newWeaponRotationVelocity, settings.SwaySmoothing); - + transform.localRotation = Quaternion.Euler(newWeaponRotation); } } -- 2.49.0 From 3420d14cdc0bc595b4e5b0eee3bb8d7b508fb4b7 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Wed, 4 May 2022 23:57:38 +0700 Subject: [PATCH 03/17] merge --- .../Timers/Greatest_map_ever_timers.json | 2 +- Assets/Prefabs/Bot.prefab | 3 +- .../Greatest_map_ever/Greatest_map_ever.unity | 29 ++++++++++- Assets/Settings/Game Settings.asset | 50 ++++++++++--------- 4 files changed, 57 insertions(+), 27 deletions(-) diff --git a/Assets/ML-Agents/Timers/Greatest_map_ever_timers.json b/Assets/ML-Agents/Timers/Greatest_map_ever_timers.json index 9a97eb3..b96488e 100644 --- a/Assets/ML-Agents/Timers/Greatest_map_ever_timers.json +++ b/Assets/ML-Agents/Timers/Greatest_map_ever_timers.json @@ -1 +1 @@ -{"count":1,"self":16.593504,"total":17.166309,"children":{"InitializeActuators":{"count":2,"self":0.0019996,"total":0.0019996,"children":null},"InitializeSensors":{"count":2,"self":0.0030004,"total":0.0030004,"children":null},"AgentSendState":{"count":612,"self":0.016998,"total":0.549809,"children":{"CollectObservations":{"count":1224,"self":0.0230042,"total":0.0230042,"children":null},"WriteActionMask":{"count":1224,"self":0.0060088,"total":0.0060088,"children":null},"RequestDecision":{"count":1224,"self":0.503798,"total":0.503798,"children":null}}},"DecideAction":{"count":612,"self":0.0089991,"total":0.0089991,"children":null},"AgentAct":{"count":612,"self":0.006997,"total":0.006997,"children":null}},"gauges":{"My Behavior.CumulativeReward":{"count":14,"max":0,"min":0,"runningAverage":0,"value":0,"weightedAverage":0}},"metadata":{"timer_format_version":"0.1.0","start_time_seconds":"1650253950","unity_version":"2019.4.35f1","command_line_arguments":"C:\\Program Files\\unityeditorfolder\\2019.4.35f1\\Editor\\Unity.exe -projectpath F:\\SigmaRiskManagment\\real shooter Git Version -useHub -hubIPC -cloudEnvironment production -licensingIpc LicenseClient-1IWpvtxiu_rvPpHhMWpzt -hubSessionId b1d8b690-be9f-11ec-92bc-6fd1276b6775 -accessToken D1AF5mitRE4Vh3s6p7rpmGuqcqvTsZJKmoGtSNfyNNs009f","communication_protocol_version":"1.5.0","com.unity.ml-agents_version":"2.0.1","scene_name":"Greatest_map_ever","end_time_seconds":"1650253967"}} \ No newline at end of file +{"count":1,"self":5.2745576,"total":5.3091238,"children":{"InitializeActuators":{"count":2,"self":0.0019998999999999998,"total":0.0019998999999999998,"children":null},"InitializeSensors":{"count":2,"self":0.0030009999999999998,"total":0.0030009999999999998,"children":null},"AgentSendState":{"count":18,"self":0.0019986,"total":0.0175794,"children":{"CollectObservations":{"count":36,"self":0.0009994,"total":0.0009994,"children":null},"WriteActionMask":{"count":36,"self":0.0005783,"total":0.0005783,"children":null},"RequestDecision":{"count":36,"self":0.0140031,"total":0.0140031,"children":null}}},"DecideAction":{"count":18,"self":0.0089876,"total":0.0089876,"children":null},"AgentAct":{"count":18,"self":0.0009983,"total":0.0009983,"children":null}},"gauges":{"My Behavior.CumulativeReward":{"count":2,"max":0,"min":0,"runningAverage":0,"value":0,"weightedAverage":0}},"metadata":{"timer_format_version":"0.1.0","start_time_seconds":"1651670021","unity_version":"2019.4.35f1","command_line_arguments":"C:\\Program Files\\unityeditorfolder\\2019.4.35f1\\Editor\\Unity.exe -projectpath F:\\SigmaRiskManagment\\real shooter Git Version -useHub -hubIPC -cloudEnvironment production -licensingIpc LicenseClient-9THNgaHTf51SEKo5URf9r -hubSessionId bf0b1ec0-c926-11ec-9066-7725c5249b8e -accessToken pLNs79-R0y77otVjQKfPq_jck-T6ncSnk21wIQRhdZw009f","communication_protocol_version":"1.5.0","com.unity.ml-agents_version":"2.0.1","scene_name":"Greatest_map_ever","end_time_seconds":"1651670026"}} \ No newline at end of file diff --git a/Assets/Prefabs/Bot.prefab b/Assets/Prefabs/Bot.prefab index eedbf35..619d624 100755 --- a/Assets/Prefabs/Bot.prefab +++ b/Assets/Prefabs/Bot.prefab @@ -143,7 +143,7 @@ MonoBehaviour: m_Name: m_EditorClassIdentifier: m_BrainParameters: - VectorObservationSize: 1 + VectorObservationSize: 9 NumStackedVectorObservations: 1 m_ActionSpec: m_NumContinuousActions: 0 @@ -195,6 +195,7 @@ MonoBehaviour: m_Name: m_EditorClassIdentifier: navMeshAgent: {fileID: 8656710265340117963} + flag: {fileID: 6818223691859422291, guid: 1685c1d9ce4ab174f95c646b1826010b, type: 3} --- !u!114 &1208561866453126566 MonoBehaviour: m_ObjectHideFlags: 0 diff --git a/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity b/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity index 41d83a6..3bcfbee 100755 --- a/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity +++ b/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity @@ -259,6 +259,8 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: + navType: 1 + PointId: 0 DeathAttr: 0 EnemiesSeen: [] --- !u!23 &140697607 @@ -353,6 +355,8 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: + navType: 1 + PointId: 0 DeathAttr: 0 EnemiesSeen: [] --- !u!23 &293522541 @@ -1315,6 +1319,8 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: + navType: 1 + PointId: 0 DeathAttr: 0 EnemiesSeen: [] --- !u!23 &1116745545 @@ -1477,7 +1483,7 @@ Mesh: m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: pb_Mesh17416 + m_Name: pb_Mesh16900 serializedVersion: 10 m_SubMeshes: - serializedVersion: 2 @@ -1763,6 +1769,8 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: + navType: 1 + PointId: 0 DeathAttr: 0 EnemiesSeen: [] --- !u!23 &1345085343 @@ -1857,6 +1865,8 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: + navType: 1 + PointId: 0 DeathAttr: 0 EnemiesSeen: [] --- !u!23 &1488699524 @@ -2077,6 +2087,8 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: + navType: 1 + PointId: 0 DeathAttr: 0 EnemiesSeen: [] --- !u!23 &1663305224 @@ -2141,6 +2153,7 @@ GameObject: - component: {fileID: 1858987090} - component: {fileID: 1858987088} - component: {fileID: 1858987085} + - component: {fileID: 1858987091} m_Layer: 0 m_Name: Game m_TagString: Untagged @@ -2253,6 +2266,18 @@ MonoBehaviour: type: 3} PlayerPrefab: {fileID: 5245491127989480125, guid: 80f6c1c85e5daed4c96c70205ed5503d, type: 3} +--- !u!114 &1858987091 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1858987083} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: bf3fe86787bfb0c4b8751fe495148ede, type: 3} + m_Name: + m_EditorClassIdentifier: --- !u!1 &1871257865 GameObject: m_ObjectHideFlags: 0 @@ -2506,6 +2531,8 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: + navType: 1 + PointId: 0 DeathAttr: 0 EnemiesSeen: [] --- !u!23 &2004854095 diff --git a/Assets/Settings/Game Settings.asset b/Assets/Settings/Game Settings.asset index f1a62c0..fc07488 100755 --- a/Assets/Settings/Game Settings.asset +++ b/Assets/Settings/Game Settings.asset @@ -12,27 +12,29 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: e2c47233b9062c84482336b145c6891b, type: 3} m_Name: Game Settings m_EditorClassIdentifier: - isTesting: 1 - timeToWin: 5 - timeOut: 1600 - movementDistance: 50 - movementSpeed: 3 - defTeamAI: 0 - atcTeamAI: 0 - numOfDefenders: 1 - numOfAttackers: 1 - hasHumanDefender: 1 - hasHumanAttacker: 0 - healthPickupAmount: 50 - armourPickupAmount: 50 - ammunitionPickupAmount: 120 - pickupsAmount: 2 - maxHealth: 0 - maxArmour: 0 - maxAmmo: 0 - GetHitChanceInDirectPoint: 0 - GetHitChanceInRunning: 0 - GetHitChanceInCover: 0 - DoDamageChanceInDirectPoint: 0 - DoDamageChanceInRunning: 0 - DoDamageChanceInCover: 0 + IsTesting: 0 + TimeToWin: 15 + TimeOut: 3600 + MovementDistance: 50 + MovementSpeed: 5 + DefTeamAI: 3 + AtcTeamAI: 3 + NumOfDefenders: 1 + NumOfAttackers: 1 + HasHumanDefender: 0 + HasHumanAttacker: 0 + HealthPickupAmount: 50 + ArmourPickupAmount: 50 + AmmunitionPickupAmount: 60 + PickupsAmount: 0 + MaxHealth: 100 + MaxArmour: 100 + MaxAmmo: 360 + ViewDistance: 100 + GetHitChanceInDirectPoint: 50 + GetHitChanceInRunning: 25 + GetHitChanceInCover: 20 + DoDamageChanceInDirectPoint: 70 + DoDamageChanceInRunning: 30 + DoDamageChanceInCover: 25 + CrouchingCoefficient: 1.4 -- 2.49.0 From e689927b730e87ecd8ba02559bde925a7df5b09d Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Thu, 5 May 2022 00:55:51 +0700 Subject: [PATCH 04/17] dadwa --- StyleCop.Cache | 7477 ++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 7477 insertions(+) create mode 100644 StyleCop.Cache diff --git a/StyleCop.Cache b/StyleCop.Cache new file mode 100644 index 0000000..edcafe4 --- /dev/null +++ b/StyleCop.Cache @@ -0,0 +1,7477 @@ + + 12 + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 01:08:14.230 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + Using directives must be sorted alphabetically by the namespaces. + 2 + False + + + The class must have a documentation header. + 5 + False + + + The field must have a documentation header. + 7 + False + + + The property must have a documentation header. + 8 + False + + + Adjacent elements must be separated by a blank line. + 8 + False + + + The property must not be placed on a single line. The opening and closing curly brackets must each be placed on their own line. + 8 + False + + + All properties must be placed after all fields. + 10 + False + + + All properties must be placed after all fields. + 11 + False + + + All properties must be placed after all fields. + 12 + False + + + All properties must be placed after all fields. + 13 + False + + + All properties must be placed after all fields. + 15 + False + + + All properties must be placed after all fields. + 16 + False + + + The call to instance must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 8 + 213 + 220 + 8 + 53 + 8 + 60 + False + + + The field must have a documentation header. + 10 + False + + + The field must have a documentation header. + 11 + False + + + The field must have a documentation header. + 12 + False + + + Variable names and private field names must start with a lower-case letter: AIPrefab. + 12 + False + + + The field must have a documentation header. + 13 + False + + + Variable names and private field names must start with a lower-case letter: PlayerPrefab. + 13 + False + + + The field must have a documentation header. + 15 + False + + + Variable names and private field names must start with a lower-case letter: Bots. + 15 + False + + + The field must have a documentation header. + 16 + False + + + Variable names and private field names must start with a lower-case letter: Player. + 16 + False + + + The method must have a documentation header. + 18 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 21 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 23 + False + + + The call to instance must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 21 + 657 + 664 + 21 + 13 + 21 + 20 + False + + + The call to instance must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 20 + 626 + 633 + 20 + 13 + 20 + 20 + False + + + The call to Destroy must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 23 + 701 + 707 + 23 + 13 + 23 + 19 + False + + + The call to gameObject must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 23 + 709 + 718 + 23 + 21 + 23 + 30 + False + + + The method must have a documentation header. + 26 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 34 + False + + + The body of the for statement must be wrapped in opening and closing curly brackets. + 37 + False + + + The body of the for statement must be wrapped in opening and closing curly brackets. + 40 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 43 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 46 + False + + + All method parameters must be placed on the same line, or each parameter must be placed on a separate line. + 37 + False + + + If the method parameters are on separate lines, the first parameter must begin on the line beneath the name of the method. + 37 + False + + + The parameter must begin on the line after the previous parameter. + 37 + False + + + All method parameters must be placed on the same line, or each parameter must be placed on a separate line. + 40 + False + + + If the method parameters are on separate lines, the first parameter must begin on the line beneath the name of the method. + 40 + False + + + The parameter must begin on the line after the previous parameter. + 40 + False + + + All method parameters must be placed on the same line, or each parameter must be placed on a separate line. + 43 + False + + + If the method parameters are on separate lines, the first parameter must begin on the line beneath the name of the method. + 43 + False + + + The parameter must begin on the line after the previous parameter. + 43 + False + + + All method parameters must be placed on the same line, or each parameter must be placed on a separate line. + 46 + False + + + If the method parameters are on separate lines, the first parameter must begin on the line beneath the name of the method. + 46 + False + + + The parameter must begin on the line after the previous parameter. + 46 + False + + + The call to InstanciateEntity must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 37 + 1298 + 1314 + 37 + 13 + 37 + 29 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 38 + 1366 + 1392 + 38 + 17 + 38 + 43 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 38 + 1410 + 1436 + 38 + 61 + 38 + 87 + False + + + The call to InstanciateEntity must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 40 + 1514 + 1530 + 40 + 13 + 40 + 29 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 41 + 1582 + 1608 + 41 + 17 + 41 + 43 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 41 + 1626 + 1652 + 41 + 61 + 41 + 87 + False + + + The call to InstanciateEntity must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 43 + 1705 + 1721 + 43 + 13 + 43 + 29 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 44 + 1772 + 1798 + 44 + 17 + 44 + 43 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 44 + 1816 + 1842 + 44 + 61 + 44 + 87 + False + + + The call to InstanciateEntity must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 46 + 1895 + 1911 + 46 + 13 + 46 + 29 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 47 + 1962 + 1988 + 47 + 17 + 47 + 43 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 47 + 2006 + 2032 + 47 + 61 + 47 + 87 + False + + + The call to ResetCharacters must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 49 + 2083 + 2097 + 49 + 37 + 49 + 51 + False + + + The method must have a documentation header. + 52 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 60 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 62 + False + + + The call to PlayerPrefab must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 55 + 2289 + 2300 + 55 + 40 + 55 + 51 + False + + + The call to AIPrefab must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 55 + 2304 + 2311 + 55 + 55 + 55 + 62 + False + + + The call to Player must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 67 + 2693 + 2698 + 67 + 13 + 67 + 18 + False + + + The call to Bots must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 73 + 2922 + 2925 + 73 + 13 + 73 + 16 + False + + + The method must have a documentation header. + 77 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 84 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 86 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 93 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 95 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 87 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 84 + 3236 + 3262 + 84 + 42 + 84 + 68 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 84 + 3280 + 3306 + 84 + 86 + 84 + 112 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 86 + 3386 + 3412 + 86 + 42 + 86 + 68 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 86 + 3430 + 3456 + 86 + 86 + 86 + 112 + False + + + The call to Bots must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 79 + 3036 + 3039 + 79 + 29 + 79 + 32 + False + + + The call to Player must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 93 + 3688 + 3693 + 93 + 17 + 93 + 22 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 93 + 3716 + 3742 + 93 + 45 + 93 + 71 + False + + + The call to spawnPointsForAttackersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 93 + 3760 + 3786 + 93 + 89 + 93 + 115 + False + + + The call to Player must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 95 + 3841 + 3846 + 95 + 17 + 95 + 22 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 95 + 3869 + 3895 + 95 + 45 + 95 + 71 + False + + + The call to spawnPointsForDefendersTeam must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 95 + 3913 + 3939 + 95 + 89 + 95 + 115 + False + + + The call to TryGetComponent<Player> must begin with the 'this.', 'base.', 'object.' or 'CharacterFactory.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 89 + 3524 + 3546 + 89 + 13 + 89 + 35 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.026 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The enum must have a documentation header. + 1 + False + + + The enumeration sub-item must have a documentation header. + 3 + False + + + The enumeration sub-item must have a documentation header. + 4 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.027 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The enum must have a documentation header. + 1 + False + + + The enumeration sub-item must have a documentation header. + 3 + False + + + The enumeration sub-item must have a documentation header. + 4 + False + + + The enumeration sub-item must have a documentation header. + 5 + False + + + The enumeration sub-item must have a documentation header. + 6 + False + + + The enumeration sub-item must have a documentation header. + 7 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.18 10:42:05.218 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 2 + False + + + Adjacent elements must be separated by a blank line. + 2 + False + + + All classes must be placed after all interfaces. + 15 + False + + + The property must have a documentation header. + 4 + False + + + All properties must be placed after all fields. + 4 + False + + + All properties must be placed after all constructors. + 4 + False + + + The field must have a documentation header. + 5 + False + + + Adjacent elements must be separated by a blank line. + 5 + False + + + Fields must be declared with private access. Use properties to expose fields. + 5 + False + + + The field must have a documentation header. + 6 + False + + + Fields must be declared with private access. Use properties to expose fields. + 6 + False + + + The constructor must have a documentation header. + 8 + False + + + The call to Condition must begin with the 'this.' prefix to indicate that the item is a member of the class. + 11 + 229 + 237 + 11 + 9 + 11 + 17 + False + + + The interface must have a documentation header. + 15 + False + + + The property must have a documentation header. + 17 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 01:00:00.568 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The code must not contain multiple blank lines in a row. + 4 + False + + + The class must have a documentation header. + 5 + False + + + The event must have a documentation header. + 7 + False + + + All events must be placed after all fields. + 11 + False + + + All events must be placed after all fields. + 35 + False + + + All events must be placed after all fields. + 48 + False + + + All events must be placed after all constructors. + 62 + False + + + The event must have a documentation header. + 8 + False + + + Adjacent elements must be separated by a blank line. + 8 + False + + + The event must have a documentation header. + 9 + False + + + Adjacent elements must be separated by a blank line. + 9 + False + + + The field must have a documentation header. + 11 + False + + + The property must have a documentation header. + 12 + False + + + Adjacent elements must be separated by a blank line. + 12 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 17 + False + + + The call to health must begin with the 'this.' prefix to indicate that the item is a member of the class. + 16 + 342 + 347 + 16 + 20 + 16 + 25 + False + + + Adjacent elements must be separated by a blank line. + 18 + False + + + The call to health must begin with the 'this.' prefix to indicate that the item is a member of the class. + 20 + 408 + 413 + 20 + 13 + 20 + 18 + False + + + The call to OnChangeHealthEvent must begin with the 'this.' prefix to indicate that the item is a member of the class. + 21 + 437 + 455 + 21 + 13 + 21 + 31 + False + + + The method must have a documentation header. + 25 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 28 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 30 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 32 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 33 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 34 + False + + + All methods must be placed after all properties. + 36 + False + + + All methods must be placed after all properties. + 49 + False + + + The call to health must begin with the 'this.' prefix to indicate that the item is a member of the class. + 27 + 558 + 563 + 27 + 13 + 27 + 18 + False + + + The call to health must begin with the 'this.' prefix to indicate that the item is a member of the class. + 29 + 612 + 617 + 29 + 18 + 29 + 23 + False + + + The call to health must begin with the 'this.' prefix to indicate that the item is a member of the class. + 31 + 666 + 671 + 31 + 18 + 31 + 23 + False + + + The field must have a documentation header. + 35 + False + + + Adjacent elements must be separated by a blank line. + 35 + False + + + The property must have a documentation header. + 36 + False + + + Adjacent elements must be separated by a blank line. + 36 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 47 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 41 + False + + + The call to armour must begin with the 'this.' prefix to indicate that the item is a member of the class. + 40 + 843 + 848 + 40 + 20 + 40 + 25 + False + + + Adjacent elements must be separated by a blank line. + 42 + False + + + The call to armour must begin with the 'this.' prefix to indicate that the item is a member of the class. + 44 + 907 + 912 + 44 + 13 + 44 + 18 + False + + + The call to OnChangeArmourEvent must begin with the 'this.' prefix to indicate that the item is a member of the class. + 45 + 936 + 954 + 45 + 13 + 45 + 31 + False + + + The field must have a documentation header. + 48 + False + + + Adjacent elements must be separated by a blank line. + 48 + False + + + The property must have a documentation header. + 49 + False + + + Adjacent elements must be separated by a blank line. + 49 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 54 + False + + + The call to ammo must begin with the 'this.' prefix to indicate that the item is a member of the class. + 53 + 1091 + 1094 + 53 + 20 + 53 + 23 + False + + + Adjacent elements must be separated by a blank line. + 55 + False + + + The call to ammo must begin with the 'this.' prefix to indicate that the item is a member of the class. + 57 + 1153 + 1156 + 57 + 13 + 57 + 16 + False + + + The call to OnChangeAmmunitionEvent must begin with the 'this.' prefix to indicate that the item is a member of the class. + 58 + 1180 + 1202 + 58 + 13 + 58 + 35 + False + + + The constructor must have a documentation header. + 62 + False + + + The call to ammo must begin with the 'this.' prefix to indicate that the item is a member of the class. + 65 + 1350 + 1353 + 65 + 9 + 65 + 12 + False + + + The call to health must begin with the 'this.' prefix to indicate that the item is a member of the class. + 66 + 1384 + 1389 + 66 + 9 + 66 + 14 + False + + + The call to armour must begin with the 'this.' prefix to indicate that the item is a member of the class. + 67 + 1422 + 1427 + 67 + 9 + 67 + 14 + False + + + The method must have a documentation header. + 70 + False + + + The call to HealthPoints must begin with the 'this.' prefix to indicate that the item is a member of the class. + 70 + 1503 + 1514 + 70 + 43 + 70 + 54 + False + + + The call to HealthPoints must begin with the 'this.' prefix to indicate that the item is a member of the class. + 70 + 1539 + 1550 + 70 + 79 + 70 + 90 + False + + + The method must have a documentation header. + 71 + False + + + Adjacent elements must be separated by a blank line. + 71 + False + + + The call to ArmourPoints must begin with the 'this.' prefix to indicate that the item is a member of the class. + 71 + 1605 + 1616 + 71 + 43 + 71 + 54 + False + + + The call to ArmourPoints must begin with the 'this.' prefix to indicate that the item is a member of the class. + 71 + 1641 + 1652 + 71 + 79 + 71 + 90 + False + + + The method must have a documentation header. + 72 + False + + + Adjacent elements must be separated by a blank line. + 72 + False + + + The call to Ammunition must begin with the 'this.' prefix to indicate that the item is a member of the class. + 72 + 1703 + 1712 + 72 + 39 + 72 + 48 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.28 18:24:37.771 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + Using directives must be sorted alphabetically by the namespaces. + 1 + False + + + System using directives must be placed before all other using directives. + 5 + False + + + The class must have a documentation header. + 8 + False + + + The code must not contain multiple blank lines in a row. + 17 + False + + + The property must have a documentation header. + 10 + False + + + All properties must be placed after all fields. + 13 + False + + + All properties must be placed after all fields. + 14 + False + + + All properties must be placed after all fields. + 15 + False + + + The property must have a documentation header. + 11 + False + + + Adjacent elements must be separated by a blank line. + 11 + False + + + The property must have a documentation header. + 12 + False + + + Adjacent elements must be separated by a blank line. + 12 + False + + + The field must have a documentation header. + 13 + False + + + Adjacent elements must be separated by a blank line. + 13 + False + + + The field must have a documentation header. + 14 + False + + + Constants must start with an upper-case letter: updateFlagPositionDelay. + 14 + False + + + The field must have a documentation header. + 15 + False + + + The method must have a documentation header. + 18 + False + + + All private methods must be placed after all public methods. + 34 + False + + + All private methods must be placed after all public methods. + 40 + False + + + All private methods must be placed after all public methods. + 47 + False + + + The call to navMeshAgent must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 20 + 555 + 566 + 20 + 9 + 20 + 20 + False + + + The call to InvokeRepeating must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 21 + 636 + 650 + 21 + 9 + 21 + 23 + False + + + The call to UpdateFlagPosition must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 21 + 659 + 676 + 21 + 32 + 21 + 49 + False + + + The method must have a documentation header. + 24 + False + + + The call to CancelInvoke must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 26 + 764 + 775 + 26 + 9 + 26 + 20 + False + + + The call to UpdateFlagPosition must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 26 + 784 + 801 + 26 + 29 + 26 + 46 + False + + + The method must have a documentation header. + 29 + False + + + The call to FlagDistance must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 31 + 870 + 881 + 31 + 9 + 31 + 20 + False + + + The call to flag must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 31 + 886 + 889 + 31 + 25 + 31 + 28 + False + + + The method must have a documentation header. + 34 + False + + + The call to goToNextNavPoint must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 37 + 1069 + 1084 + 37 + 9 + 37 + 24 + False + + + The method must have a documentation header. + 40 + False + + + method names begin with an upper-case letter: getPointsCandidate. + 40 + False + + + The method must have a documentation header. + 47 + False + + + method names begin with an upper-case letter: goToNextNavPoint. + 47 + False + + + The call to navMeshAgent must begin with the 'this.', 'base.', 'object.' or 'MovementController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 48 + 1492 + 1503 + 48 + 9 + 48 + 20 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.18 09:03:31.781 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The enum must have a documentation header. + 1 + False + + + The enumeration sub-item must have a documentation header. + 3 + False + + + The enumeration sub-item must have a documentation header. + 4 + False + + + The enumeration sub-item must have a documentation header. + 5 + False + + + The interface must have a documentation header. + 8 + False + + + Interface names must start with the capital letter 'I': NPC_BaseState. + 8 + False + + + The property must have a documentation header. + 10 + False + + + The property must have a documentation header. + 11 + False + + + Adjacent elements must be separated by a blank line. + 11 + False + + + The property must have a documentation header. + 12 + False + + + Adjacent elements must be separated by a blank line. + 12 + False + + + The property must have a documentation header. + 13 + False + + + Adjacent elements must be separated by a blank line. + 13 + False + + + The property must have a documentation header. + 14 + False + + + Adjacent elements must be separated by a blank line. + 14 + False + + + The property must have a documentation header. + 15 + False + + + Adjacent elements must be separated by a blank line. + 15 + False + + + The class must have a documentation header. + 18 + False + + + The property must have a documentation header. + 20 + False + + + The property must have a documentation header. + 21 + False + + + Adjacent elements must be separated by a blank line. + 21 + False + + + The property must have a documentation header. + 22 + False + + + Adjacent elements must be separated by a blank line. + 22 + False + + + The property must have a documentation header. + 23 + False + + + Adjacent elements must be separated by a blank line. + 23 + False + + + The property must have a documentation header. + 24 + False + + + Adjacent elements must be separated by a blank line. + 24 + False + + + The property must have a documentation header. + 25 + False + + + Adjacent elements must be separated by a blank line. + 25 + False + + + The class must have a documentation header. + 28 + False + + + A C# document may only contain a single class at the root level unless all of the classes are partial and are of the same type. + 28 + False + + + The property must have a documentation header. + 30 + False + + + The property must have a documentation header. + 31 + False + + + Adjacent elements must be separated by a blank line. + 31 + False + + + The property must have a documentation header. + 32 + False + + + Adjacent elements must be separated by a blank line. + 32 + False + + + The property must have a documentation header. + 33 + False + + + Adjacent elements must be separated by a blank line. + 33 + False + + + The property must have a documentation header. + 34 + False + + + Adjacent elements must be separated by a blank line. + 34 + False + + + The property must have a documentation header. + 35 + False + + + Adjacent elements must be separated by a blank line. + 35 + False + + + The class must have a documentation header. + 38 + False + + + The property must have a documentation header. + 40 + False + + + The property must have a documentation header. + 41 + False + + + Adjacent elements must be separated by a blank line. + 41 + False + + + The property must have a documentation header. + 42 + False + + + Adjacent elements must be separated by a blank line. + 42 + False + + + The property must have a documentation header. + 43 + False + + + Adjacent elements must be separated by a blank line. + 43 + False + + + The property must have a documentation header. + 44 + False + + + Adjacent elements must be separated by a blank line. + 44 + False + + + The property must have a documentation header. + 45 + False + + + Adjacent elements must be separated by a blank line. + 45 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.18 10:43:46.328 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 4 + False + + + The field must have a documentation header. + 7 + False + + + Fields must be declared with private access. Use properties to expose fields. + 7 + False + + + The field must have a documentation header. + 8 + False + + + Fields must be declared with private access. Use properties to expose fields. + 8 + False + + + The property must have a documentation header. + 10 + False + + + All properties must be placed after all events. + 23 + False + + + The call to PlayerCharacter must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 10 + 232 + 246 + 10 + 38 + 10 + 52 + False + + + The method must have a documentation header. + 12 + False + + + All private methods must be placed after all public methods. + 18 + False + + + All private methods must be placed after all public methods. + 24 + False + + + The call to PlayerCharacter must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 14 + 293 + 307 + 14 + 9 + 14 + 23 + False + + + The call to Condition must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 15 + 337 + 345 + 15 + 9 + 15 + 17 + False + + + The call to PlayerCharacter must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 15 + 349 + 363 + 15 + 21 + 15 + 35 + False + + + The method must have a documentation header. + 18 + False + + + The call to Condition must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 20 + 435 + 443 + 20 + 9 + 20 + 17 + False + + + The event must have a documentation header. + 23 + False + + + The method must have a documentation header. + 24 + False + + + Adjacent elements must be separated by a blank line. + 24 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 31 + False + + + Insert parentheses within the arithmetic expression to declare the operator precedence. + 27 + False + + + The call to PlayerCharacter must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 26 + 587 + 601 + 26 + 9 + 26 + 23 + False + + + The call to Condition must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 27 + 660 + 668 + 27 + 9 + 27 + 17 + False + + + The call to Condition must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 27 + 713 + 721 + 27 + 62 + 27 + 70 + False + + + The call to Condition must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 28 + 756 + 764 + 28 + 9 + 28 + 17 + False + + + The call to OnKilledEvent must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 31 + 884 + 896 + 31 + 13 + 31 + 25 + False + + + The call to Condition must begin with the 'this.', 'base.', 'object.' or 'Player.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 30 + 836 + 844 + 30 + 13 + 30 + 21 + False + + + The method must have a documentation header. + 34 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.030 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 9 + False + + + An opening curly bracket must not be followed by a blank line. + 10 + False + + + A closing curly bracket must not be preceded by a blank line. + 259 + False + + + class names begin with an upper-case letter: scr_CharacterController. + 9 + False + + + The field must have a documentation header. + 12 + False + + + All private fields must be placed after all public fields. + 16 + False + + + All private fields must be placed after all public fields. + 22 + False + + + All private fields must be placed after all public fields. + 23 + False + + + All private fields must be placed after all public fields. + 26 + False + + + All private fields must be placed after all public fields. + 28 + False + + + All private fields must be placed after all public fields. + 29 + False + + + All private fields must be placed after all public fields. + 30 + False + + + All private fields must be placed after all public fields. + 33 + False + + + All private fields must be placed after all public fields. + 34 + False + + + All private fields must be placed after all public fields. + 37 + False + + + All private fields must be placed after all public fields. + 41 + False + + + All private fields must be placed after all public fields. + 42 + False + + + All private fields must be placed after all public fields. + 43 + False + + + All private fields must be placed after all public fields. + 44 + False + + + All private fields must be placed after all public fields. + 45 + False + + + All private fields must be placed after all public fields. + 56 + False + + + The field must have a documentation header. + 13 + False + + + The field must have a documentation header. + 14 + False + + + Field names must not contain underscores. + 14 + False + + + The field must have a documentation header. + 16 + False + + + Fields must be declared with private access. Use properties to expose fields. + 16 + False + + + Field names must not contain underscores. + 16 + False + + + Public and internal fields must start with an upper-case letter: input_View. + 16 + False + + + The field must have a documentation header. + 18 + False + + + The field must have a documentation header. + 19 + False + + + The field must have a documentation header. + 22 + False + + + Fields must be declared with private access. Use properties to expose fields. + 22 + False + + + Public and internal fields must start with an upper-case letter: cameraHolder. + 22 + False + + + The field must have a documentation header. + 23 + False + + + Fields must be declared with private access. Use properties to expose fields. + 23 + False + + + Public and internal fields must start with an upper-case letter: feetTransform. + 23 + False + + + The field must have a documentation header. + 26 + False + + + Fields must be declared with private access. Use properties to expose fields. + 26 + False + + + Public and internal fields must start with an upper-case letter: playerSettings. + 26 + False + + + The field must have a documentation header. + 28 + False + + + Fields must be declared with private access. Use properties to expose fields. + 28 + False + + + The field must have a documentation header. + 29 + False + + + Fields must be declared with private access. Use properties to expose fields. + 29 + False + + + The field must have a documentation header. + 30 + False + + + Fields must be declared with private access. Use properties to expose fields. + 30 + False + + + Public and internal fields must start with an upper-case letter: playerMask. + 30 + False + + + The field must have a documentation header. + 33 + False + + + Fields must be declared with private access. Use properties to expose fields. + 33 + False + + + Public and internal fields must start with an upper-case letter: gravityAmount. + 33 + False + + + The field must have a documentation header. + 34 + False + + + Fields must be declared with private access. Use properties to expose fields. + 34 + False + + + Public and internal fields must start with an upper-case letter: gravityMin. + 34 + False + + + The field must have a documentation header. + 35 + False + + + The field must have a documentation header. + 37 + False + + + Fields must be declared with private access. Use properties to expose fields. + 37 + False + + + Public and internal fields must start with an upper-case letter: jumpingForce. + 37 + False + + + The field must have a documentation header. + 38 + False + + + The field must have a documentation header. + 41 + False + + + Fields must be declared with private access. Use properties to expose fields. + 41 + False + + + Public and internal fields must start with an upper-case letter: playerStance. + 41 + False + + + The field must have a documentation header. + 42 + False + + + Fields must be declared with private access. Use properties to expose fields. + 42 + False + + + Public and internal fields must start with an upper-case letter: playerStanceSmoothing. + 42 + False + + + The field must have a documentation header. + 43 + False + + + Fields must be declared with private access. Use properties to expose fields. + 43 + False + + + Public and internal fields must start with an upper-case letter: playerStandStance. + 43 + False + + + The field must have a documentation header. + 44 + False + + + Fields must be declared with private access. Use properties to expose fields. + 44 + False + + + Public and internal fields must start with an upper-case letter: playerCrouchStance. + 44 + False + + + The field must have a documentation header. + 45 + False + + + Fields must be declared with private access. Use properties to expose fields. + 45 + False + + + Public and internal fields must start with an upper-case letter: playerProneStance. + 45 + False + + + The field must have a documentation header. + 46 + False + + + The field must have a documentation header. + 48 + False + + + The field must have a documentation header. + 49 + False + + + The field must have a documentation header. + 51 + False + + + The field must have a documentation header. + 53 + False + + + The field must have a documentation header. + 54 + False + + + The field must have a documentation header. + 56 + False + + + Fields must be declared with private access. Use properties to expose fields. + 56 + False + + + Public and internal fields must start with an upper-case letter: currentWeapon. + 56 + False + + + The method must have a documentation header. + 57 + False + + + Adjacent elements must be separated by a blank line. + 57 + False + + + A closing curly bracket must not be preceded by a blank line. + 84 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 59 + 1597 + 1608 + 59 + 9 + 59 + 20 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 61 + 1643 + 1654 + 61 + 9 + 61 + 20 + False + + + The call to input_Movement must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 61 + 1693 + 1706 + 61 + 59 + 61 + 72 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 62 + 1743 + 1754 + 62 + 9 + 62 + 20 + False + + + The call to input_View must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 62 + 1789 + 1798 + 62 + 55 + 62 + 64 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 63 + 1835 + 1846 + 63 + 9 + 63 + 20 + False + + + The call to Jump must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 63 + 1881 + 1884 + 63 + 55 + 63 + 58 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 65 + 1908 + 1919 + 65 + 9 + 65 + 20 + False + + + The call to Crouch must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 65 + 1956 + 1961 + 65 + 57 + 65 + 62 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 66 + 1975 + 1986 + 66 + 9 + 66 + 20 + False + + + The call to Prone must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 66 + 2022 + 2026 + 66 + 56 + 66 + 60 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 68 + 2050 + 2061 + 68 + 9 + 68 + 20 + False + + + The call to ToggleSprint must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 68 + 2098 + 2109 + 68 + 57 + 68 + 68 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 69 + 2123 + 2134 + 69 + 9 + 69 + 20 + False + + + The call to StopSprint must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 69 + 2179 + 2188 + 69 + 65 + 69 + 74 + False + + + The call to defaultInput must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 71 + 2210 + 2221 + 71 + 9 + 71 + 20 + False + + + The call to newCameraRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 73 + 2244 + 2260 + 73 + 9 + 73 + 25 + False + + + The call to cameraHolder must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 73 + 2264 + 2275 + 73 + 29 + 73 + 40 + False + + + The call to newCharacterRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 74 + 2313 + 2332 + 74 + 9 + 74 + 28 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 75 + 2382 + 2400 + 75 + 9 + 75 + 27 + False + + + The call to GetComponent<CharacterController> must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 75 + 2404 + 2436 + 75 + 31 + 75 + 63 + False + + + The call to cameraHeight must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 77 + 2452 + 2463 + 77 + 9 + 77 + 20 + False + + + The call to cameraHolder must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 77 + 2467 + 2478 + 77 + 24 + 77 + 35 + False + + + The call to currentWeapon must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 81 + 2551 + 2563 + 81 + 13 + 81 + 25 + False + + + The call to currentWeapon must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 79 + 2512 + 2524 + 79 + 13 + 79 + 25 + False + + + The method must have a documentation header. + 86 + False + + + The call to CalculateView must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 88 + 2656 + 2668 + 88 + 9 + 88 + 21 + False + + + The call to CalculateMovement must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 89 + 2682 + 2698 + 89 + 9 + 89 + 25 + False + + + The call to CalculateJump must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 90 + 2712 + 2724 + 90 + 9 + 90 + 21 + False + + + The call to CalculateCameraHeight must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 91 + 2738 + 2758 + 91 + 9 + 91 + 29 + False + + + The method must have a documentation header. + 94 + False + + + The call to newCharacterRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 96 + 2822 + 2841 + 96 + 9 + 96 + 28 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 96 + 2848 + 2861 + 96 + 35 + 96 + 48 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 96 + 2883 + 2896 + 96 + 70 + 96 + 83 + False + + + The call to input_View must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 96 + 2915 + 2924 + 96 + 102 + 96 + 111 + False + + + The call to input_View must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 96 + 2930 + 2939 + 96 + 117 + 96 + 126 + False + + + The call to newCharacterRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 97 + 3014 + 3033 + 97 + 52 + 97 + 71 + False + + + The call to newCameraRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 99 + 3056 + 3072 + 99 + 9 + 99 + 25 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 99 + 3079 + 3092 + 99 + 32 + 99 + 45 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 99 + 3114 + 3127 + 99 + 67 + 99 + 80 + False + + + The call to input_View must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 99 + 3145 + 3154 + 99 + 98 + 99 + 107 + False + + + The call to input_View must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 99 + 3161 + 3170 + 99 + 114 + 99 + 123 + False + + + The call to newCameraRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 100 + 3202 + 3218 + 100 + 9 + 100 + 25 + False + + + The call to newCameraRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 100 + 3236 + 3252 + 100 + 43 + 100 + 59 + False + + + The call to ViewClampYMin must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 100 + 3257 + 3269 + 100 + 64 + 100 + 76 + False + + + The call to ViewClampYMax must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 100 + 3272 + 3284 + 100 + 79 + 100 + 91 + False + + + The call to cameraHolder must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 102 + 3307 + 3318 + 102 + 9 + 102 + 20 + False + + + The call to newCameraRotation must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 102 + 3353 + 3369 + 102 + 55 + 102 + 71 + False + + + The method must have a documentation header. + 105 + False + + + Variable names and private field names must start with a lower-case letter: MovementSpeed. + 147 + False + + + All method parameters must be placed on the same line, or each parameter must be placed on a separate line. + 142 + False + + + If the method parameters are on separate lines, the first parameter must begin on the line beneath the name of the method. + 142 + False + + + The parameter must begin on the line after the previous parameter. + 145 + False + + + All method parameters must be placed on the same line, or each parameter must be placed on a separate line. + 143 + False + + + If the method parameters are on separate lines, the first parameter must begin on the line beneath the name of the method. + 143 + False + + + The parameter must begin on the line after the previous parameter. + 144 + False + + + The call to isSprinting must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 109 + 3490 + 3500 + 109 + 13 + 109 + 23 + False + + + The call to input_Movement must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 107 + 3440 + 3453 + 107 + 13 + 107 + 26 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 112 + 3561 + 3574 + 112 + 29 + 112 + 42 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 113 + 3628 + 3641 + 113 + 31 + 113 + 44 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 117 + 3731 + 3744 + 117 + 29 + 117 + 42 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 118 + 3798 + 3811 + 118 + 31 + 118 + 44 + False + + + The call to isSprinting must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 115 + 3678 + 3688 + 115 + 13 + 115 + 23 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 124 + 3946 + 3959 + 124 + 13 + 124 + 26 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 124 + 3977 + 3990 + 124 + 44 + 124 + 57 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 122 + 3890 + 3908 + 122 + 14 + 122 + 32 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 128 + 4103 + 4116 + 128 + 13 + 128 + 26 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 128 + 4134 + 4147 + 128 + 44 + 128 + 57 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 126 + 4042 + 4053 + 126 + 17 + 126 + 28 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 132 + 4259 + 4272 + 132 + 13 + 132 + 26 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 132 + 4290 + 4303 + 132 + 44 + 132 + 57 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 130 + 4199 + 4210 + 130 + 17 + 130 + 28 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 136 + 4374 + 4387 + 136 + 13 + 136 + 26 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 139 + 4455 + 4468 + 139 + 26 + 139 + 39 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 140 + 4513 + 4526 + 140 + 28 + 140 + 41 + False + + + The call to newMovementSpeed must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 142 + 4562 + 4577 + 142 + 9 + 142 + 24 + False + + + The call to newMovementSpeed must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 142 + 4600 + 4615 + 142 + 47 + 142 + 62 + False + + + The call to input_Movement must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 143 + 4656 + 4669 + 143 + 38 + 143 + 51 + False + + + The call to input_Movement must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 144 + 4721 + 4734 + 144 + 30 + 144 + 43 + False + + + The call to newMovementSpeedVelocity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 145 + 4774 + 4797 + 145 + 17 + 145 + 40 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 145 + 4800 + 4818 + 145 + 43 + 145 + 61 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 145 + 4833 + 4846 + 145 + 76 + 145 + 89 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 145 + 4868 + 4881 + 145 + 111 + 145 + 124 + False + + + The call to newMovementSpeed must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 147 + 4970 + 4985 + 147 + 58 + 147 + 73 + False + + + The call to playerGravity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 151 + 5056 + 5068 + 151 + 13 + 151 + 25 + False + + + The call to gravityAmount must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 151 + 5073 + 5085 + 151 + 30 + 151 + 42 + False + + + The call to playerGravity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 149 + 5004 + 5016 + 149 + 13 + 149 + 25 + False + + + The call to gravityMin must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 149 + 5020 + 5029 + 149 + 29 + 149 + 38 + False + + + The call to playerGravity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 156 + 5212 + 5224 + 156 + 13 + 156 + 25 + False + + + The call to playerGravity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 154 + 5131 + 5143 + 154 + 13 + 154 + 25 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 154 + 5156 + 5174 + 154 + 38 + 154 + 56 + False + + + The call to playerGravity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 159 + 5276 + 5288 + 159 + 28 + 159 + 40 + False + + + The call to jumpingForce must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 160 + 5317 + 5328 + 160 + 26 + 160 + 37 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 162 + 5367 + 5385 + 162 + 9 + 162 + 27 + False + + + The spacing around the keyword 'if' is invalid. + 126 + 4039 + 4040 + 126 + 14 + 126 + 15 + False + + + The spacing around the keyword 'if' is invalid. + 130 + 4196 + 4197 + 130 + 14 + 130 + 15 + False + + + The method must have a documentation header. + 165 + False + + + The call to jumpingForce must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 167 + 5467 + 5478 + 167 + 9 + 167 + 20 + False + + + The call to jumpingForce must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 167 + 5501 + 5512 + 167 + 43 + 167 + 54 + False + + + The call to jumpingForceVelocity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 167 + 5533 + 5552 + 167 + 75 + 167 + 94 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 167 + 5555 + 5568 + 167 + 97 + 167 + 110 + False + + + The method must have a documentation header. + 170 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 186 + False + + + The call to playerStandStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 172 + 5673 + 5689 + 172 + 28 + 172 + 44 + False + + + The call to playerCrouchStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 176 + 5796 + 5813 + 176 + 28 + 176 + 45 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 174 + 5720 + 5731 + 174 + 13 + 174 + 24 + False + + + The call to playerProneStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 180 + 5933 + 5949 + 180 + 28 + 180 + 44 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 178 + 5858 + 5869 + 178 + 18 + 178 + 29 + False + + + The call to cameraHeight must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 183 + 5995 + 6006 + 183 + 9 + 183 + 20 + False + + + The call to cameraHolder must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 183 + 6027 + 6038 + 183 + 41 + 183 + 52 + False + + + The call to cameraHeightVelocity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 183 + 6075 + 6094 + 183 + 89 + 183 + 108 + False + + + The call to playerStanceSmoothing must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 183 + 6097 + 6117 + 183 + 111 + 183 + 131 + False + + + The call to cameraHolder must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 185 + 6132 + 6143 + 185 + 9 + 185 + 20 + False + + + The call to cameraHolder must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 185 + 6173 + 6184 + 185 + 50 + 185 + 61 + False + + + The call to cameraHeight must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 185 + 6203 + 6214 + 185 + 80 + 185 + 91 + False + + + The call to cameraHolder must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 185 + 6217 + 6228 + 185 + 94 + 185 + 105 + False + + + The method must have a documentation header. + 187 + False + + + Adjacent elements must be separated by a blank line. + 187 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 199 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 189 + 6301 + 6319 + 189 + 14 + 189 + 32 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 189 + 6335 + 6346 + 189 + 48 + 189 + 59 + False + + + The call to StanceCheck must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 196 + 6502 + 6512 + 196 + 17 + 196 + 27 + False + + + The call to playerStandStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 196 + 6514 + 6530 + 196 + 29 + 196 + 45 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 200 + 6624 + 6635 + 200 + 13 + 200 + 24 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 194 + 6437 + 6448 + 194 + 13 + 194 + 24 + False + + + The call to jumpingForce must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 205 + 6727 + 6738 + 205 + 9 + 205 + 20 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 205 + 6755 + 6768 + 205 + 37 + 205 + 50 + False + + + The call to playerGravity must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 206 + 6794 + 6806 + 206 + 9 + 206 + 21 + False + + + The method must have a documentation header. + 209 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 216 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 219 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 223 + False + + + The call to StanceCheck must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 213 + 6934 + 6944 + 213 + 17 + 213 + 27 + False + + + The call to playerStandStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 213 + 6946 + 6962 + 213 + 29 + 213 + 45 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 217 + 7056 + 7067 + 217 + 13 + 217 + 24 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 211 + 6869 + 6880 + 211 + 13 + 211 + 24 + False + + + The call to StanceCheck must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 220 + 7136 + 7146 + 220 + 13 + 220 + 23 + False + + + The call to playerCrouchStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 220 + 7148 + 7165 + 220 + 25 + 220 + 42 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 224 + 7243 + 7254 + 224 + 9 + 224 + 20 + False + + + The method must have a documentation header. + 227 + False + + + The call to playerStance must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 229 + 7330 + 7341 + 229 + 9 + 229 + 20 + False + + + The method must have a documentation header. + 232 + False + + + The code must not contain multiple blank lines in a row. + 237 + False + + + The call to feetTransform must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 234 + 7469 + 7481 + 234 + 33 + 234 + 45 + False + + + The call to feetTransform must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 234 + 7495 + 7507 + 234 + 59 + 234 + 71 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 234 + 7522 + 7540 + 234 + 86 + 234 + 104 + False + + + The call to stanceCheckErrorMargin must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 234 + 7551 + 7572 + 234 + 115 + 234 + 136 + False + + + The call to feetTransform must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 234 + 7575 + 7587 + 234 + 139 + 234 + 151 + False + + + The call to feetTransform must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 235 + 7633 + 7645 + 235 + 31 + 235 + 43 + False + + + The call to feetTransform must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 235 + 7659 + 7671 + 235 + 57 + 235 + 69 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 235 + 7686 + 7704 + 235 + 84 + 235 + 102 + False + + + The call to stanceCheckErrorMargin must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 235 + 7715 + 7736 + 235 + 113 + 235 + 134 + False + + + The call to feetTransform must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 235 + 7759 + 7771 + 235 + 157 + 235 + 169 + False + + + The call to characterController must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 238 + 7855 + 7873 + 238 + 49 + 238 + 67 + False + + + The call to playerMask must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 238 + 7883 + 7892 + 238 + 77 + 238 + 86 + False + + + The method must have a documentation header. + 241 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 247 + False + + + The call to isSprinting must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 245 + 8008 + 8018 + 245 + 13 + 245 + 23 + False + + + The call to input_Movement must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 243 + 7958 + 7971 + 243 + 13 + 243 + 26 + False + + + The call to isSprinting must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 248 + 8070 + 8080 + 248 + 9 + 248 + 19 + False + + + The call to isSprinting must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 248 + 8085 + 8095 + 248 + 24 + 248 + 34 + False + + + The method must have a documentation header. + 251 + False + + + The call to isSprinting must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 255 + 8216 + 8226 + 255 + 13 + 255 + 23 + False + + + The call to playerSettings must begin with the 'this.', 'base.', 'object.' or 'scr_CharacterController.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 253 + 8162 + 8175 + 253 + 13 + 253 + 26 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.031 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 4 + False + + + class names begin with an upper-case letter: scr_Models. + 4 + False + + + The enum must have a documentation header. + 8 + False + + + The enumeration sub-item must have a documentation header. + 10 + False + + + The enumeration sub-item must have a documentation header. + 11 + False + + + The enumeration sub-item must have a documentation header. + 12 + False + + + The class must have a documentation header. + 16 + False + + + The field must have a documentation header. + 19 + False + + + Fields must be declared with private access. Use properties to expose fields. + 19 + False + + + The field must have a documentation header. + 20 + False + + + Fields must be declared with private access. Use properties to expose fields. + 20 + False + + + The field must have a documentation header. + 22 + False + + + Fields must be declared with private access. Use properties to expose fields. + 22 + False + + + The field must have a documentation header. + 23 + False + + + Fields must be declared with private access. Use properties to expose fields. + 23 + False + + + The field must have a documentation header. + 26 + False + + + Fields must be declared with private access. Use properties to expose fields. + 26 + False + + + The field must have a documentation header. + 27 + False + + + Fields must be declared with private access. Use properties to expose fields. + 27 + False + + + The field must have a documentation header. + 30 + False + + + Fields must be declared with private access. Use properties to expose fields. + 30 + False + + + The field must have a documentation header. + 31 + False + + + Fields must be declared with private access. Use properties to expose fields. + 31 + False + + + The field must have a documentation header. + 34 + False + + + Fields must be declared with private access. Use properties to expose fields. + 34 + False + + + The field must have a documentation header. + 35 + False + + + Fields must be declared with private access. Use properties to expose fields. + 35 + False + + + The field must have a documentation header. + 36 + False + + + Fields must be declared with private access. Use properties to expose fields. + 36 + False + + + The field must have a documentation header. + 39 + False + + + Fields must be declared with private access. Use properties to expose fields. + 39 + False + + + The field must have a documentation header. + 40 + False + + + Fields must be declared with private access. Use properties to expose fields. + 40 + False + + + The field must have a documentation header. + 41 + False + + + Fields must be declared with private access. Use properties to expose fields. + 41 + False + + + The field must have a documentation header. + 44 + False + + + Fields must be declared with private access. Use properties to expose fields. + 44 + False + + + The field must have a documentation header. + 45 + False + + + Fields must be declared with private access. Use properties to expose fields. + 45 + False + + + The field must have a documentation header. + 46 + False + + + Fields must be declared with private access. Use properties to expose fields. + 46 + False + + + The field must have a documentation header. + 47 + False + + + Fields must be declared with private access. Use properties to expose fields. + 47 + False + + + The class must have a documentation header. + 51 + False + + + The field must have a documentation header. + 53 + False + + + Fields must be declared with private access. Use properties to expose fields. + 53 + False + + + The field must have a documentation header. + 54 + False + + + Fields must be declared with private access. Use properties to expose fields. + 54 + False + + + The class must have a documentation header. + 62 + False + + + The field must have a documentation header. + 65 + False + + + Fields must be declared with private access. Use properties to expose fields. + 65 + False + + + The field must have a documentation header. + 66 + False + + + Fields must be declared with private access. Use properties to expose fields. + 66 + False + + + The field must have a documentation header. + 67 + False + + + Fields must be declared with private access. Use properties to expose fields. + 67 + False + + + The field must have a documentation header. + 68 + False + + + Fields must be declared with private access. Use properties to expose fields. + 68 + False + + + The field must have a documentation header. + 69 + False + + + Fields must be declared with private access. Use properties to expose fields. + 69 + False + + + The field must have a documentation header. + 70 + False + + + Fields must be declared with private access. Use properties to expose fields. + 70 + False + + + The field must have a documentation header. + 71 + False + + + Fields must be declared with private access. Use properties to expose fields. + 71 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 01:02:47.068 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + System using directives must be placed before all other using directives. + 3 + False + + + The class must have a documentation header. + 5 + False + + + The field must have a documentation header. + 7 + False + + + The property must have a documentation header. + 8 + False + + + Adjacent elements must be separated by a blank line. + 8 + False + + + The property must not be placed on a single line. The opening and closing curly brackets must each be placed on their own line. + 8 + False + + + All properties must be placed after all fields. + 10 + False + + + All properties must be placed after all fields. + 11 + False + + + All properties must be placed after all events. + 98 + False + + + The field must have a documentation header. + 10 + False + + + Variable names and private field names must start with a lower-case letter: DefendersTeam. + 10 + False + + + The field must have a documentation header. + 11 + False + + + Variable names and private field names must start with a lower-case letter: AttackersTeam. + 11 + False + + + The method must have a documentation header. + 13 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 16 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 18 + False + + + All private methods must be placed after all public methods. + 39 + False + + + All private methods must be placed after all public methods. + 54 + False + + + All private methods must be placed after all public methods. + 68 + False + + + The call to Destroy must begin with the 'this.', 'base.', 'object.' or 'GameManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 18 + 536 + 542 + 18 + 13 + 18 + 19 + False + + + The call to gameObject must begin with the 'this.', 'base.', 'object.' or 'GameManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 18 + 544 + 553 + 18 + 21 + 18 + 30 + False + + + The method must have a documentation header. + 21 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 33 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 35 + False + + + The call to ResetScene must begin with the 'this.', 'base.', 'object.' or 'GameManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 23 + 647 + 656 + 23 + 48 + 23 + 57 + False + + + The call to flagCaptured must begin with the 'this.', 'base.', 'object.' or 'GameManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 25 + 706 + 717 + 25 + 45 + 25 + 56 + False + + + The call to timeOut must begin with the 'this.', 'base.', 'object.' or 'GameManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 26 + 762 + 768 + 26 + 42 + 26 + 48 + False + + + The method must have a documentation header. + 39 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 43 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 45 + False + + + The body of the foreach statement must be wrapped in opening and closing curly brackets. + 49 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 50 + False + + + The method must have a documentation header. + 54 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 58 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 60 + False + + + The body of the foreach statement must be wrapped in opening and closing curly brackets. + 63 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 64 + False + + + The method must have a documentation header. + 68 + False + + + The method must have a documentation header. + 71 + False + + + method names begin with an upper-case letter: flagCaptured. + 71 + False + + + The spacing around the keyword 'switch' is invalid. + 73 + 2522 + 2527 + 73 + 9 + 73 + 14 + False + + + The method must have a documentation header. + 87 + False + + + method names begin with an upper-case letter: timeOut. + 87 + False + + + The method must have a documentation header. + 92 + False + + + The call to flagCaptured must begin with the 'this.', 'base.', 'object.' or 'GameManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 94 + 3031 + 3042 + 94 + 45 + 94 + 56 + False + + + The call to timeOut must begin with the 'this.', 'base.', 'object.' or 'GameManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 95 + 3087 + 3093 + 95 + 42 + 95 + 48 + False + + + The event must have a documentation header. + 98 + False + + + The method must have a documentation header. + 99 + False + + + Adjacent elements must be separated by a blank line. + 99 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.033 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 3 + False + + + The event must have a documentation header. + 5 + False + + + event names begin with an upper-case letter: onCaptureFlag. + 5 + False + + + The method must have a documentation header. + 7 + False + + + All methods must be placed after all events. + 13 + False + + + The event must have a documentation header. + 13 + False + + + event names begin with an upper-case letter: onTimeLeft. + 13 + False + + + The method must have a documentation header. + 14 + False + + + Adjacent elements must be separated by a blank line. + 14 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.16 19:49:10.120 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 4 + False + + + The property must have a documentation header. + 6 + False + + + property names begin with an upper-case letter: navPoints. + 6 + False + + + The method must have a documentation header. + 7 + False + + + Adjacent elements must be separated by a blank line. + 7 + False + + + A line may only contain a single statement. + 15 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.16 00:20:04.403 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 5 + False + + + The field must have a documentation header. + 7 + False + + + The property must have a documentation header. + 8 + False + + + Adjacent elements must be separated by a blank line. + 8 + False + + + The property must not be placed on a single line. The opening and closing curly brackets must each be placed on their own line. + 8 + False + + + The property must have a documentation header. + 10 + False + + + The method must have a documentation header. + 11 + False + + + Adjacent elements must be separated by a blank line. + 11 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 23 + False + + + The method must have an access modifier. + 11 + False + + + The call to Destroy must begin with the 'this.', 'base.', 'object.' or 'TimeManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 21 + 529 + 535 + 21 + 13 + 21 + 19 + False + + + The call to gameObject must begin with the 'this.', 'base.', 'object.' or 'TimeManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 21 + 537 + 546 + 21 + 21 + 21 + 30 + False + + + The method must have a documentation header. + 24 + False + + + Adjacent elements must be separated by a blank line. + 24 + False + + + The method must have an access modifier. + 24 + False + + + The call to CurrentTime must begin with the 'this.', 'base.', 'object.' or 'TimeManager.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 26 + 603 + 613 + 26 + 9 + 26 + 19 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.16 20:17:09.487 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 5 + False + + + A single-line comment must be preceded by a blank line or another single-line comment, or must be the first item in its scope. To ignore this error when commenting out a line of code, begin the comment with '////' rather than '//'. + 14 + False + + + A single-line comment must not be followed by a blank line. To ignore this error when commenting out a line of code, begin the comment with '////' rather than '//'. + 14 + False + + + The comment must start with a single space. To ignore this error when commenting out a line of code, begin the comment with '////' rather than '//'. + 14 + 391 + 414 + 14 + 5 + 14 + 28 + False + + + The property must have a documentation header. + 7 + False + + + property names begin with an upper-case letter: position. + 7 + False + + + All properties must be placed after all fields. + 11 + False + + + All properties must be placed after all fields. + 12 + False + + + All properties must be placed after all fields. + 13 + False + + + The property must have a documentation header. + 8 + False + + + Adjacent elements must be separated by a blank line. + 8 + False + + + The field must have a documentation header. + 11 + False + + + Fields must be declared with private access. Use properties to expose fields. + 11 + False + + + The field must have a documentation header. + 12 + False + + + Fields must be declared with private access. Use properties to expose fields. + 12 + False + + + The field must have a documentation header. + 13 + False + + + Fields must be declared with private access. Use properties to expose fields. + 13 + False + + + The method must have a documentation header. + 16 + False + + + The call to FlagDistance must begin with the 'this.', 'base.', 'object.' or 'NavPoint.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 18 + 460 + 471 + 18 + 9 + 18 + 20 + False + + + The call to position must begin with the 'this.', 'base.', 'object.' or 'NavPoint.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 18 + 538 + 545 + 18 + 87 + 18 + 94 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 01:00:02.216 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 4 + False + + + The spacing around the symbol '=' is invalid. + 3 + False + + + The field must have a documentation header. + 6 + False + + + Fields must be declared with private access. Use properties to expose fields. + 6 + False + + + The field must have a documentation header. + 8 + False + + + Fields must be declared with private access. Use properties to expose fields. + 8 + False + + + The field must have a documentation header. + 9 + False + + + Fields must be declared with private access. Use properties to expose fields. + 9 + False + + + The field must have a documentation header. + 12 + False + + + Fields must be declared with private access. Use properties to expose fields. + 12 + False + + + The field must have a documentation header. + 13 + False + + + Fields must be declared with private access. Use properties to expose fields. + 13 + False + + + The field must have a documentation header. + 15 + False + + + Fields must be declared with private access. Use properties to expose fields. + 15 + False + + + The field must have a documentation header. + 16 + False + + + Fields must be declared with private access. Use properties to expose fields. + 16 + False + + + The field must have a documentation header. + 17 + False + + + Fields must be declared with private access. Use properties to expose fields. + 17 + False + + + The field must have a documentation header. + 18 + False + + + Fields must be declared with private access. Use properties to expose fields. + 18 + False + + + The field must have a documentation header. + 19 + False + + + Fields must be declared with private access. Use properties to expose fields. + 19 + False + + + The field must have a documentation header. + 20 + False + + + Fields must be declared with private access. Use properties to expose fields. + 20 + False + + + The field must have a documentation header. + 22 + False + + + Fields must be declared with private access. Use properties to expose fields. + 22 + False + + + The field must have a documentation header. + 23 + False + + + Fields must be declared with private access. Use properties to expose fields. + 23 + False + + + The field must have a documentation header. + 24 + False + + + Fields must be declared with private access. Use properties to expose fields. + 24 + False + + + The field must have a documentation header. + 25 + False + + + Fields must be declared with private access. Use properties to expose fields. + 25 + False + + + The field must have a documentation header. + 27 + False + + + Fields must be declared with private access. Use properties to expose fields. + 27 + False + + + The field must have a documentation header. + 28 + False + + + Fields must be declared with private access. Use properties to expose fields. + 28 + False + + + The field must have a documentation header. + 29 + False + + + Fields must be declared with private access. Use properties to expose fields. + 29 + False + + + The field must have a documentation header. + 31 + False + + + Fields must be declared with private access. Use properties to expose fields. + 31 + False + + + The field must have a documentation header. + 33 + False + + + Fields must be declared with private access. Use properties to expose fields. + 33 + False + + + The field must have a documentation header. + 34 + False + + + Fields must be declared with private access. Use properties to expose fields. + 34 + False + + + The field must have a documentation header. + 35 + False + + + Fields must be declared with private access. Use properties to expose fields. + 35 + False + + + The field must have a documentation header. + 36 + False + + + Fields must be declared with private access. Use properties to expose fields. + 36 + False + + + The field must have a documentation header. + 37 + False + + + Fields must be declared with private access. Use properties to expose fields. + 37 + False + + + The field must have a documentation header. + 38 + False + + + Fields must be declared with private access. Use properties to expose fields. + 38 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 23:29:50.464 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 5 + False + + + The field must have a documentation header. + 7 + False + + + The property must have a documentation header. + 8 + False + + + Adjacent elements must be separated by a blank line. + 8 + False + + + The property must not be placed on a single line. The opening and closing curly brackets must each be placed on their own line. + 8 + False + + + All properties must be placed after all fields. + 15 + False + + + The method must have a documentation header. + 10 + False + + + All methods must be placed after all properties. + 16 + False + + + The field must have a documentation header. + 15 + False + + + The property must have a documentation header. + 16 + False + + + Adjacent elements must be separated by a blank line. + 16 + False + + + The property must not be placed on a single line. The opening and closing curly brackets must each be placed on their own line. + 16 + False + + + The call to gameSettings must begin with the 'this.', 'base.', 'object.' or 'SettingsReader.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 16 + 419 + 430 + 16 + 48 + 16 + 59 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.040 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 3 + False + + + The method must have a documentation header. + 5 + False + + + An opening curly bracket must not be followed by a blank line. + 6 + False + + + A closing curly bracket must not be preceded by a blank line. + 8 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 00:59:43.057 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 5 + False + + + The property must have a documentation header. + 7 + False + + + property names begin with an upper-case letter: type. + 7 + False + + + The method must have a documentation header. + 9 + False + + + The call to PickObject must begin with the 'this.', 'base.', 'object.' or 'AmmoPickUp.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 11 + 252 + 261 + 11 + 9 + 11 + 18 + False + + + The method must have a documentation header. + 14 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 00:59:39.476 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 5 + False + + + The property must have a documentation header. + 7 + False + + + property names begin with an upper-case letter: type. + 7 + False + + + The method must have a documentation header. + 9 + False + + + The call to PickObject must begin with the 'this.', 'base.', 'object.' or 'ArmourPickUp.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 11 + 250 + 259 + 11 + 9 + 11 + 18 + False + + + The method must have a documentation header. + 14 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 00:59:35.943 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 5 + False + + + The property must have a documentation header. + 7 + False + + + property names begin with an upper-case letter: type. + 7 + False + + + The method must have a documentation header. + 9 + False + + + The call to PickObject must begin with the 'this.', 'base.', 'object.' or 'HealthPickUp.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 11 + 250 + 259 + 11 + 9 + 11 + 18 + False + + + The method must have a documentation header. + 14 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.043 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The interface must have a documentation header. + 3 + False + + + Adjacent elements must be separated by a blank line. + 3 + False + + + The property must have a documentation header. + 5 + False + + + property names begin with an upper-case letter: type. + 5 + False + + + The method must have a documentation header. + 6 + False + + + Adjacent elements must be separated by a blank line. + 6 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 00:59:47.515 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 7 + False + + + The field must have a documentation header. + 9 + False + + + The property must have a documentation header. + 10 + False + + + Adjacent elements must be separated by a blank line. + 10 + False + + + The property must not be placed on a single line. The opening and closing curly brackets must each be placed on their own line. + 10 + False + + + All properties must be placed after all fields. + 12 + False + + + All properties must be placed after all fields. + 13 + False + + + All properties must be placed after all fields. + 14 + False + + + All properties must be placed after all fields. + 15 + False + + + All properties must be placed after all fields. + 17 + False + + + The call to instance must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 10 + 255 + 262 + 10 + 50 + 10 + 57 + False + + + The field must have a documentation header. + 12 + False + + + The field must have a documentation header. + 13 + False + + + The field must have a documentation header. + 14 + False + + + The field must have a documentation header. + 15 + False + + + The field must have a documentation header. + 17 + False + + + The method must have a documentation header. + 19 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 22 + False + + + The body of the else statement must be wrapped in opening and closing curly brackets. + 24 + False + + + The call to instance must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 22 + 612 + 619 + 22 + 13 + 22 + 20 + False + + + The call to instance must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 21 + 581 + 588 + 21 + 13 + 21 + 20 + False + + + The call to Destroy must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 24 + 656 + 662 + 24 + 13 + 24 + 19 + False + + + The call to gameObject must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 24 + 664 + 673 + 24 + 21 + 24 + 30 + False + + + The method must have a documentation header. + 27 + False + + + The body of the for statement must be wrapped in opening and closing curly brackets. + 32 + False + + + The body of the for statement must be wrapped in opening and closing curly brackets. + 34 + False + + + The body of the foreach statement must be wrapped in opening and closing curly brackets. + 40 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 38 + False + + + The call to pickups must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 29 + 728 + 734 + 29 + 9 + 29 + 15 + False + + + The call to pickups must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 32 + 890 + 896 + 32 + 13 + 32 + 19 + False + + + The call to healthPrefab must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 32 + 925 + 936 + 32 + 48 + 32 + 59 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 32 + 939 + 949 + 32 + 62 + 32 + 72 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 32 + 967 + 977 + 32 + 90 + 32 + 100 + False + + + The call to pickups must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 34 + 1085 + 1091 + 34 + 13 + 34 + 19 + False + + + The call to armourPrefab must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 34 + 1120 + 1131 + 34 + 48 + 34 + 59 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 34 + 1134 + 1144 + 34 + 62 + 34 + 72 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 34 + 1162 + 1172 + 34 + 90 + 34 + 100 + False + + + The call to pickups must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 37 + 1291 + 1297 + 37 + 13 + 37 + 19 + False + + + The call to ammoPrefab must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 37 + 1326 + 1335 + 37 + 48 + 37 + 57 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 37 + 1338 + 1348 + 37 + 60 + 37 + 70 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 37 + 1366 + 1376 + 37 + 88 + 37 + 98 + False + + + The call to pickups must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 39 + 1473 + 1479 + 39 + 33 + 39 + 39 + False + + + The call to StartCoroutine must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 42 + 1531 + 1544 + 42 + 9 + 42 + 22 + False + + + The call to SpawnNewPickUps must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 42 + 1546 + 1560 + 42 + 24 + 42 + 38 + False + + + The method must have a documentation header. + 45 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 58 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 55 + 1898 + 1908 + 55 + 47 + 55 + 57 + False + + + The call to spawnPoints must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 55 + 1926 + 1936 + 55 + 75 + 55 + 85 + False + + + The call to IsDisableCheck must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 50 + 1703 + 1716 + 50 + 16 + 50 + 29 + False + + + The spacing around the keyword 'while' is invalid. + 47 + 1634 + 1638 + 47 + 9 + 47 + 13 + False + + + The spacing around the keyword 'if' is invalid. + 50 + 1700 + 1701 + 50 + 13 + 50 + 14 + False + + + The code contains multiple spaces in a row. Only one space is needed. + 59 + 2062 + 2063 + 59 + 29 + 59 + 30 + False + + + The method must have a documentation header. + 63 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 72 + False + + + The call to pickups must begin with the 'this.', 'base.', 'object.' or 'PickUpSpawner.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 65 + 2196 + 2202 + 65 + 29 + 65 + 35 + False + + + The spacing around the keyword 'foreach' is invalid. + 65 + 2176 + 2182 + 65 + 9 + 65 + 15 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.044 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The enum must have a documentation header. + 1 + False + + + The enumeration sub-item must have a documentation header. + 3 + False + + + The enumeration sub-item must have a documentation header. + 4 + False + + + The enumeration sub-item must have a documentation header. + 5 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 22:38:30.693 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The code must not contain blank lines at the end of the file. + 4 + False + + + The code must not contain multiple blank lines in a row. + 4 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.046 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The enum must have a documentation header. + 1 + False + + + The enumeration sub-item must have a documentation header. + 3 + False + + + The enumeration sub-item must have a documentation header. + 4 + False + + + The enumeration sub-item must have a documentation header. + 5 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.24 00:44:55.312 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 1 + False + + + The method must have a documentation header. + 3 + False + + + Field names must not start with an underscore. + 3 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 16:12:15.866 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 11 + False + + + The field must have a documentation header. + 13 + False + + + The field must have a documentation header. + 14 + False + + + The field must have a documentation header. + 15 + False + + + The field must have a documentation header. + 16 + False + + + The field must have a documentation header. + 17 + False + + + The field must have a documentation header. + 18 + False + + + The field must have a documentation header. + 19 + False + + + The field must have a documentation header. + 20 + False + + + The field must have a documentation header. + 21 + False + + + The field must have a documentation header. + 23 + False + + + The property must have a documentation header. + 26 + False + + + The property must have a documentation header. + 31 + False + + + The indexer must have a documentation header. + 36 + False + + + The indexer must have a documentation header. + 47 + False + + + The constructor must have a documentation header. + 60 + False + + + The constructor must have a documentation header. + 65 + False + + + The constructor must have a documentation header. + 70 + False + + + The constructor must have a documentation header. + 75 + False + + + The constructor must have a documentation header. + 85 + False + + + The constructor must have a documentation header. + 90 + False + + + The method must have a documentation header. + 100 + False + + + The method must have a documentation header. + 122 + False + + + The method must have a documentation header. + 127 + False + + + The method must have a documentation header. + 146 + False + + + The method must have a documentation header. + 151 + False + + + The method must have a documentation header. + 190 + False + + + The method must have a documentation header. + 195 + False + + + The method must have a documentation header. + 226 + False + + + The method must have a documentation header. + 281 + False + + + The method must have a documentation header. + 297 + False + + + The method must have a documentation header. + 314 + False + + + The class must have a documentation header. + 325 + False + + + The field must have a documentation header. + 327 + False + + + The method must have a documentation header. + 403 + False + + + The method must have a documentation header. + 420 + False + + + The method must have a documentation header. + 439 + False + + + The property must have a documentation header. + 450 + False + + + The property must have a documentation header. + 455 + False + + + The method must have a documentation header. + 460 + False + + + The method must have a documentation header. + 465 + False + + + The method must have a documentation header. + 472 + False + + + The property must have a documentation header. + 490 + False + + + The method must have a documentation header. + 495 + False + + + The method must have a documentation header. + 500 + False + + + The method must have a documentation header. + 505 + False + + + The method must have a documentation header. + 510 + False + + + + + UNITY_2019_4_35;UNITY_2019_4;UNITY_2019;UNITY_5_3_OR_NEWER;UNITY_5_4_OR_NEWER;UNITY_5_5_OR_NEWER;UNITY_5_6_OR_NEWER;UNITY_2017_1_OR_NEWER;UNITY_2017_2_OR_NEWER;UNITY_2017_3_OR_NEWER;UNITY_2017_4_OR_NEWER;UNITY_2018_1_OR_NEWER;UNITY_2018_2_OR_NEWER;UNITY_2018_3_OR_NEWER;UNITY_2018_4_OR_NEWER;UNITY_2019_1_OR_NEWER;UNITY_2019_2_OR_NEWER;UNITY_2019_3_OR_NEWER;UNITY_2019_4_OR_NEWER;PLATFORM_ARCH_64;UNITY_64;UNITY_INCLUDE_TESTS;ENABLE_AR;ENABLE_AUDIO;ENABLE_CACHING;ENABLE_CLOTH;ENABLE_EVENT_QUEUE;ENABLE_MICROPHONE;ENABLE_MULTIPLE_DISPLAYS;ENABLE_PHYSICS;ENABLE_TEXTURE_STREAMING;ENABLE_UNET;ENABLE_LZMA;ENABLE_UNITYEVENTS;ENABLE_VR;ENABLE_WEBCAM;ENABLE_UNITYWEBREQUEST;ENABLE_WWW;ENABLE_CLOUD_SERVICES;ENABLE_CLOUD_SERVICES_COLLAB;ENABLE_CLOUD_SERVICES_COLLAB_SOFTLOCKS;ENABLE_CLOUD_SERVICES_ADS;ENABLE_CLOUD_SERVICES_USE_WEBREQUEST;ENABLE_CLOUD_SERVICES_CRASH_REPORTING;ENABLE_CLOUD_SERVICES_PURCHASING;ENABLE_CLOUD_SERVICES_ANALYTICS;ENABLE_CLOUD_SERVICES_UNET;ENABLE_CLOUD_SERVICES_BUILD;ENABLE_CLOUD_LICENSE;ENABLE_EDITOR_HUB_LICENSE;ENABLE_WEBSOCKET_CLIENT;ENABLE_DIRECTOR_AUDIO;ENABLE_DIRECTOR_TEXTURE;ENABLE_MANAGED_JOBS;ENABLE_MANAGED_TRANSFORM_JOBS;ENABLE_MANAGED_ANIMATION_JOBS;ENABLE_MANAGED_AUDIO_JOBS;INCLUDE_DYNAMIC_GI;ENABLE_MONO_BDWGC;ENABLE_SCRIPTING_GC_WBARRIERS;PLATFORM_SUPPORTS_MONO;RENDER_SOFTWARE_CURSOR;ENABLE_VIDEO;PLATFORM_STANDALONE;PLATFORM_STANDALONE_WIN;UNITY_STANDALONE_WIN;UNITY_STANDALONE;ENABLE_RUNTIME_GI;ENABLE_MOVIES;ENABLE_NETWORK;ENABLE_CRUNCH_TEXTURE_COMPRESSION;ENABLE_OUT_OF_PROCESS_CRASH_HANDLER;ENABLE_CLUSTER_SYNC;ENABLE_CLUSTERINPUT;GFXDEVICE_WAITFOREVENT_MESSAGEPUMP;ENABLE_WEBSOCKET_HOST;ENABLE_MONO;NET_4_6;ENABLE_PROFILER;DEBUG;TRACE;UNITY_ASSERTIONS;UNITY_EDITOR;UNITY_EDITOR_64;UNITY_EDITOR_WIN;ENABLE_UNITY_COLLECTIONS_CHECKS;ENABLE_BURST_AOT;UNITY_TEAM_LICENSE;ENABLE_CUSTOM_RENDER_TEXTURE;ENABLE_DIRECTOR;ENABLE_LOCALIZATION;ENABLE_SPRITES;ENABLE_TERRAIN;ENABLE_TILEMAP;ENABLE_TIMELINE;ENABLE_INPUT_SYSTEM;ENABLE_LEGACY_INPUT_MANAGER;CSHARP_7_OR_LATER;CSHARP_7_3_OR_NEWER + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.28 18:30:19.041 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + A syntax error has been discovered in file F:\SigmaRiskManagment\real shooter Git Version\Assets\Scripts\Character\NPC.cs on line 87. + 87 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.28 18:30:19.035 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 5 + False + + + The property must have a documentation header. + 8 + False + + + All properties must be placed after all fields. + 10 + False + + + All properties must be placed after all fields. + 11 + False + + + All properties must be placed after all fields. + 14 + False + + + The property must have a documentation header. + 9 + False + + + Adjacent elements must be separated by a blank line. + 9 + False + + + The field must have a documentation header. + 10 + False + + + Adjacent elements must be separated by a blank line. + 10 + False + + + The field must have a documentation header. + 11 + False + + + The property must have a documentation header. + 12 + False + + + Adjacent elements must be separated by a blank line. + 12 + False + + + The call to occupDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 12 + 362 + 375 + 12 + 33 + 12 + 46 + False + + + The call to occupAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 12 + 384 + 397 + 12 + 55 + 12 + 68 + False + + + The spacing around the symbol '>' is invalid. + 12 + 376 + 376 + 12 + 47 + 12 + 47 + False + + + The spacing around the symbol '>' is invalid. + 12 + False + + + The spacing around the symbol '>' is invalid. + 12 + 398 + 398 + 12 + 69 + 12 + 69 + False + + + The property must have a documentation header. + 13 + False + + + Adjacent elements must be separated by a blank line. + 13 + False + + + The call to occupDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 13 + 435 + 448 + 13 + 32 + 13 + 45 + False + + + The call to occupAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 13 + 460 + 473 + 13 + 57 + 13 + 70 + False + + + The field must have a documentation header. + 14 + False + + + Adjacent elements must be separated by a blank line. + 14 + False + + + The method must have a documentation header. + 16 + False + + + An opening curly bracket must not be followed by a blank line. + 17 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 24 + False + + + The call to timeForWin must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 19 + 559 + 568 + 19 + 9 + 19 + 18 + False + + + The call to TimeStayAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 20 + 628 + 644 + 20 + 9 + 20 + 25 + False + + + The call to TimeStayDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 21 + 660 + 676 + 21 + 9 + 21 + 25 + False + + + The call to occupAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 22 + 692 + 705 + 22 + 9 + 22 + 22 + False + + + The call to occupDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 23 + 721 + 734 + 23 + 9 + 23 + 22 + False + + + The method must have a documentation header. + 25 + False + + + Adjacent elements must be separated by a blank line. + 25 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 39 + False + + + The call to occupDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 30 + 889 + 902 + 30 + 17 + 30 + 30 + False + + + The call to occupAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 33 + 978 + 991 + 33 + 17 + 33 + 30 + False + + + The spacing around the keyword 'switch' is invalid. + 27 + 813 + 818 + 27 + 9 + 27 + 14 + False + + + The method must have a documentation header. + 40 + False + + + Adjacent elements must be separated by a blank line. + 40 + False + + + Statements or elements wrapped in curly brackets must be followed by a blank line. + 54 + False + + + The call to occupDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 45 + 1287 + 1300 + 45 + 17 + 45 + 30 + False + + + The call to occupAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 48 + 1376 + 1389 + 48 + 17 + 48 + 30 + False + + + The method must have a documentation header. + 55 + False + + + Adjacent elements must be separated by a blank line. + 55 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 67 + False + + + The body of the if statement must be wrapped in opening and closing curly brackets. + 73 + False + + + The call to TimeStayAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 59 + 1641 + 1657 + 59 + 13 + 59 + 29 + False + + + The call to TimeStayDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 60 + 1677 + 1693 + 60 + 13 + 60 + 29 + False + + + The call to IsOccupBoth must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 57 + 1590 + 1600 + 57 + 13 + 57 + 23 + False + + + The call to IsNotOccup must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 57 + 1605 + 1614 + 57 + 28 + 57 + 37 + False + + + The call to TimeStayAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 65 + 1794 + 1810 + 65 + 13 + 65 + 29 + False + + + The call to TimeStayAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 66 + 1848 + 1864 + 66 + 17 + 66 + 33 + False + + + The call to timeForWin must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 66 + 1868 + 1877 + 66 + 37 + 66 + 46 + False + + + The call to occupAttackers must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 63 + 1750 + 1763 + 63 + 18 + 63 + 31 + False + + + The call to TimeStayDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 71 + 1999 + 2015 + 71 + 13 + 71 + 29 + False + + + The call to TimeStayDefenders must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 72 + 2053 + 2069 + 72 + 17 + 72 + 33 + False + + + The call to timeForWin must begin with the 'this.', 'base.', 'object.' or 'FlagZone.' or 'MonoBehaviour.' prefix to indicate the intended method call. + 72 + 2073 + 2082 + 72 + 37 + 72 + 46 + False + + + + + + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.710 + 2022.04.12 15:21:15.050 + 2022.04.28 18:27:01.731 + 2022.04.28 18:27:01.731 + -2051395988 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + 2022.04.28 18:27:01.731 + 0 + + + + The file has no header, the header Xml is invalid, or the header is not located at the top of the file. + 1 + False + + + The class must have a documentation header. + 4 + False + + + + \ No newline at end of file -- 2.49.0 From 690f1e253602c6fded329ceca14cc25d3b4ce70f Mon Sep 17 00:00:00 2001 From: Andrey Gumirov Date: Thu, 5 May 2022 01:08:58 +0700 Subject: [PATCH 05/17] test --- 1 | 0 1 file changed, 0 insertions(+), 0 deletions(-) create mode 100644 1 diff --git a/1 b/1 new file mode 100644 index 0000000..e69de29 -- 2.49.0 From 351fa8ee12c499a0374be71a192a5f288611f467 Mon Sep 17 00:00:00 2001 From: Andrey Gumirov Date: Thu, 5 May 2022 01:10:04 +0700 Subject: [PATCH 06/17] test 1 --- 1 | 1 + 1 file changed, 1 insertion(+) diff --git a/1 b/1 index e69de29..8b13789 100644 --- a/1 +++ b/1 @@ -0,0 +1 @@ + -- 2.49.0 From cfb7b71a8a61ae44419070002cfd73049c278d93 Mon Sep 17 00:00:00 2001 From: Andrey Gumirov Date: Thu, 5 May 2022 15:37:51 +0700 Subject: [PATCH 07/17] Fixed merge errors --- .../Scripts/Character/CharacterCondition.cs | 10 ----- .../Scripts/Character/MovementController.cs | 25 ----------- Assets/Scripts/Character/NPC.cs | 45 ------------------- Assets/Scripts/Misc/NavPoint.cs | 9 ---- 4 files changed, 89 deletions(-) diff --git a/Assets/Scripts/Character/CharacterCondition.cs b/Assets/Scripts/Character/CharacterCondition.cs index 1efa341..de60740 100755 --- a/Assets/Scripts/Character/CharacterCondition.cs +++ b/Assets/Scripts/Character/CharacterCondition.cs @@ -9,22 +9,12 @@ public class CharacterCondition public event Action OnChangeAmmunitionEvent; private int health; -<<<<<<< HEAD public int HealthPoints { get { return health; } -======= - - public int HealthPoints - { - get - { - return health; - } ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 private set { health = value; diff --git a/Assets/Scripts/Character/MovementController.cs b/Assets/Scripts/Character/MovementController.cs index 3ade88a..0022357 100644 --- a/Assets/Scripts/Character/MovementController.cs +++ b/Assets/Scripts/Character/MovementController.cs @@ -6,18 +6,11 @@ using UnityEngine.AI; [RequireComponent(typeof(NavMeshAgent))] public class MovementController : MonoBehaviour { -<<<<<<< HEAD public int PointStartID { get; set; } public int PointEndID { get; private set; } public float FlagDistance { get; private set; } private const float updateFlagPositionDelay = 5; private const float updateReachedDestinationDelay = 5; -======= - public NavPoint currentPosition { get; set; } - private Dictionary navPoints = new Dictionary(); - - [SerializeField] private NavMeshAgent navMeshAgent; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 [SerializeField] private NavMeshAgent navMeshAgent; [SerializeField] private GameObject flag; @@ -28,7 +21,6 @@ public class MovementController : MonoBehaviour private void Awake() { -<<<<<<< HEAD navMeshAgent.speed = SettingsReader.Instance.GetSettings.MovementSpeed; idNavPointDict = MapManager.IDToNavPoint; InvokeRepeating(nameof(UpdateFlagPosition), 0, updateFlagPositionDelay); @@ -43,12 +35,6 @@ public class MovementController : MonoBehaviour private void UpdateFlagPosition() { FlagDistance = (flag.transform.position - gameObject.transform.position).magnitude; -======= - navMeshAgent.speed = SettingsReader.Instance.GetSettings.movementSpeed; - foreach (var np in MapManager.navPoints) { - navPoints[np.PointId] = np; - } ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } public void MoveToPointById(int id) @@ -62,27 +48,16 @@ public class MovementController : MonoBehaviour } public void MoveToRandomPoint() -<<<<<<< HEAD { Debug.Log(MapManager.NavPoints == null); GoToNextNavPoint(MapManager.NavPoints[Random.Range(0, MapManager.NavPoints.Count)]); -======= - { - // Debug.Log(MapManager.navPoints == null); - goToNextNavPoint(MapManager.navPoints[Random.Range(0, MapManager.navPoints.Count)]); ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } public List GetPointsCandidate() { -<<<<<<< HEAD return MapManager.NavPoints .Where(point => (idNavPointDict[PointStartID].Position - point.Position).magnitude < SettingsReader.Instance.GetSettings.MovementDistance) -======= - return MapManager.navPoints - .Where(point => (currentPosition.position - point.position).magnitude < SettingsReader.Instance.GetSettings.movementDistance) ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 .ToList(); } diff --git a/Assets/Scripts/Character/NPC.cs b/Assets/Scripts/Character/NPC.cs index 932379c..d47e377 100644 --- a/Assets/Scripts/Character/NPC.cs +++ b/Assets/Scripts/Character/NPC.cs @@ -70,14 +70,7 @@ public class NPC : Agent, ICharacter public override void CollectObservations(VectorSensor sensor) { -<<<<<<< HEAD var candidates = moveController.GetPointsCandidate(); -======= - sensor.AddObservation(Condition.HealthPoints); - sensor.AddObservation(Condition.ArmourPoints); - sensor.AddObservation(Condition.Ammunition); - sensor.AddObservation((int) NPC_State.State); ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 //common sensors sensor.AddObservation(GameManager.IsHaveSeenByEnemy(AgentCharacter.Team.GetOppositeTeam(), @@ -100,7 +93,6 @@ public class NPC : Agent, ICharacter //point sensors foreach (var point in candidates) { -<<<<<<< HEAD bufferSensor.AppendObservation(new float[] { point.DeathAttr, (int)point.navType, @@ -112,35 +104,6 @@ public class NPC : Agent, ICharacter GameManager.IsHaveSeenByEnemy(AgentCharacter.Team.GetOppositeTeam(), point.Position).ToInt() }); -======= - var parray = new float[] - { - //1 position in navpointId - (float) moveController.currentPosition.PointId, - //2 distance to flag - moveController.currentPosition.FlagDistance, - //3 death count in point - moveController.currentPosition.DeathAttr, - //4 flagEnemyDistance - GameManager.IsCloserToFlagFromNextNavPoint(point, transform.position) == true ? 1 : 0, - //5 EnemyVsNavPointDistance - GameManager.IsCloserToEnemyThanToNextNavPoint(point, transform.position, AgentCharacter.Team) == true - ? 1 - : 0 - }; - // var _parray = string.Join(" ", parray); - // Debug.Log("OBS: " + _parray); - bufferSensor.AppendObservation(parray); - } - } - - public override void Heuristic(in ActionBuffers actionsOut) - { - var discreteActionsOut = actionsOut.DiscreteActions; - if (Input.GetKeyDown(KeyCode.W)) - { - discreteActionsOut[0] = 1; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } } @@ -149,7 +112,6 @@ public class NPC : Agent, ICharacter var result = actions.DiscreteActions; if (result[0] == 0) { -<<<<<<< HEAD if (navPointIdDict[moveController.PointStartID].navType != NavPointType.Cover) return; NpcState = CoverState; @@ -185,13 +147,6 @@ public class NPC : Agent, ICharacter case 1: moveController.ReturnToStartPoint(); NpcState = RunningState; break; default: throw new ArgumentException("Undefined Action recieved"); } -======= - moveController.MoveToRandomPoint(); - NPC_State = RunningState; - } else if (actions.DiscreteActions[0] == 2) - { - moveController.MoveToPointById(actions.DiscreteActions[1]); ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } } #endregion diff --git a/Assets/Scripts/Misc/NavPoint.cs b/Assets/Scripts/Misc/NavPoint.cs index 473d7fa..a70fd61 100755 --- a/Assets/Scripts/Misc/NavPoint.cs +++ b/Assets/Scripts/Misc/NavPoint.cs @@ -17,21 +17,12 @@ public class NavPoint : MonoBehaviour public NavPointType navType = NavPointType.Direction; [HideInInspector] -<<<<<<< HEAD public int PointId = 0; -======= - public int PointId; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 public float DeathAttr = 0; public List EnemiesSeen = new List(); private void Start() { -<<<<<<< HEAD FlagDistance = (GameObject.FindGameObjectWithTag("Flag").transform.position - Position).magnitude; -======= - PointId = GetInstanceID(); - FlagDistance = (GameObject.FindGameObjectWithTag("Flag").transform.position - position).magnitude; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } } -- 2.49.0 From f3fe5ea4cd8ef270fbfc6c1be90b25b6576f5a93 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Thu, 5 May 2022 16:26:23 +0700 Subject: [PATCH 08/17] commit 665 --- Assets/Prefabs/Player.prefab.meta | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Assets/Prefabs/Player.prefab.meta b/Assets/Prefabs/Player.prefab.meta index 9180d60..1fa6659 100755 --- a/Assets/Prefabs/Player.prefab.meta +++ b/Assets/Prefabs/Player.prefab.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: 99bb17a23a489624baeaf337f91a4f84 +guid: 80f6c1c85e5daed4c96c70205ed5503d PrefabImporter: externalObjects: {} userData: -- 2.49.0 From ba1b350c6283b0ebb3d6e48bf7de66db8f6503a6 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Thu, 5 May 2022 16:49:21 +0700 Subject: [PATCH 09/17] resolved merge conflicts --- .../Scripts/Character/CharacterCondition.cs | 9 ---- .../Scripts/Character/MovementController.cs | 26 ----------- Assets/Scripts/Character/NPC.cs | 45 ------------------- Assets/Scripts/Misc/NavPoint.cs | 9 ---- 4 files changed, 89 deletions(-) diff --git a/Assets/Scripts/Character/CharacterCondition.cs b/Assets/Scripts/Character/CharacterCondition.cs index 1efa341..8d7d5c5 100755 --- a/Assets/Scripts/Character/CharacterCondition.cs +++ b/Assets/Scripts/Character/CharacterCondition.cs @@ -9,14 +9,6 @@ public class CharacterCondition public event Action OnChangeAmmunitionEvent; private int health; -<<<<<<< HEAD - public int HealthPoints - { - get - { - return health; - } -======= public int HealthPoints { @@ -24,7 +16,6 @@ public class CharacterCondition { return health; } ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 private set { health = value; diff --git a/Assets/Scripts/Character/MovementController.cs b/Assets/Scripts/Character/MovementController.cs index 3ade88a..b1849ce 100644 --- a/Assets/Scripts/Character/MovementController.cs +++ b/Assets/Scripts/Character/MovementController.cs @@ -6,18 +6,11 @@ using UnityEngine.AI; [RequireComponent(typeof(NavMeshAgent))] public class MovementController : MonoBehaviour { -<<<<<<< HEAD public int PointStartID { get; set; } public int PointEndID { get; private set; } public float FlagDistance { get; private set; } private const float updateFlagPositionDelay = 5; private const float updateReachedDestinationDelay = 5; -======= - public NavPoint currentPosition { get; set; } - private Dictionary navPoints = new Dictionary(); - - [SerializeField] private NavMeshAgent navMeshAgent; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 [SerializeField] private NavMeshAgent navMeshAgent; [SerializeField] private GameObject flag; @@ -28,7 +21,6 @@ public class MovementController : MonoBehaviour private void Awake() { -<<<<<<< HEAD navMeshAgent.speed = SettingsReader.Instance.GetSettings.MovementSpeed; idNavPointDict = MapManager.IDToNavPoint; InvokeRepeating(nameof(UpdateFlagPosition), 0, updateFlagPositionDelay); @@ -43,12 +35,6 @@ public class MovementController : MonoBehaviour private void UpdateFlagPosition() { FlagDistance = (flag.transform.position - gameObject.transform.position).magnitude; -======= - navMeshAgent.speed = SettingsReader.Instance.GetSettings.movementSpeed; - foreach (var np in MapManager.navPoints) { - navPoints[np.PointId] = np; - } ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } public void MoveToPointById(int id) @@ -62,27 +48,15 @@ public class MovementController : MonoBehaviour } public void MoveToRandomPoint() -<<<<<<< HEAD { - Debug.Log(MapManager.NavPoints == null); GoToNextNavPoint(MapManager.NavPoints[Random.Range(0, MapManager.NavPoints.Count)]); -======= - { - // Debug.Log(MapManager.navPoints == null); - goToNextNavPoint(MapManager.navPoints[Random.Range(0, MapManager.navPoints.Count)]); ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } public List GetPointsCandidate() { -<<<<<<< HEAD return MapManager.NavPoints .Where(point => (idNavPointDict[PointStartID].Position - point.Position).magnitude < SettingsReader.Instance.GetSettings.MovementDistance) -======= - return MapManager.navPoints - .Where(point => (currentPosition.position - point.position).magnitude < SettingsReader.Instance.GetSettings.movementDistance) ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 .ToList(); } diff --git a/Assets/Scripts/Character/NPC.cs b/Assets/Scripts/Character/NPC.cs index 932379c..d47e377 100644 --- a/Assets/Scripts/Character/NPC.cs +++ b/Assets/Scripts/Character/NPC.cs @@ -70,14 +70,7 @@ public class NPC : Agent, ICharacter public override void CollectObservations(VectorSensor sensor) { -<<<<<<< HEAD var candidates = moveController.GetPointsCandidate(); -======= - sensor.AddObservation(Condition.HealthPoints); - sensor.AddObservation(Condition.ArmourPoints); - sensor.AddObservation(Condition.Ammunition); - sensor.AddObservation((int) NPC_State.State); ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 //common sensors sensor.AddObservation(GameManager.IsHaveSeenByEnemy(AgentCharacter.Team.GetOppositeTeam(), @@ -100,7 +93,6 @@ public class NPC : Agent, ICharacter //point sensors foreach (var point in candidates) { -<<<<<<< HEAD bufferSensor.AppendObservation(new float[] { point.DeathAttr, (int)point.navType, @@ -112,35 +104,6 @@ public class NPC : Agent, ICharacter GameManager.IsHaveSeenByEnemy(AgentCharacter.Team.GetOppositeTeam(), point.Position).ToInt() }); -======= - var parray = new float[] - { - //1 position in navpointId - (float) moveController.currentPosition.PointId, - //2 distance to flag - moveController.currentPosition.FlagDistance, - //3 death count in point - moveController.currentPosition.DeathAttr, - //4 flagEnemyDistance - GameManager.IsCloserToFlagFromNextNavPoint(point, transform.position) == true ? 1 : 0, - //5 EnemyVsNavPointDistance - GameManager.IsCloserToEnemyThanToNextNavPoint(point, transform.position, AgentCharacter.Team) == true - ? 1 - : 0 - }; - // var _parray = string.Join(" ", parray); - // Debug.Log("OBS: " + _parray); - bufferSensor.AppendObservation(parray); - } - } - - public override void Heuristic(in ActionBuffers actionsOut) - { - var discreteActionsOut = actionsOut.DiscreteActions; - if (Input.GetKeyDown(KeyCode.W)) - { - discreteActionsOut[0] = 1; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } } @@ -149,7 +112,6 @@ public class NPC : Agent, ICharacter var result = actions.DiscreteActions; if (result[0] == 0) { -<<<<<<< HEAD if (navPointIdDict[moveController.PointStartID].navType != NavPointType.Cover) return; NpcState = CoverState; @@ -185,13 +147,6 @@ public class NPC : Agent, ICharacter case 1: moveController.ReturnToStartPoint(); NpcState = RunningState; break; default: throw new ArgumentException("Undefined Action recieved"); } -======= - moveController.MoveToRandomPoint(); - NPC_State = RunningState; - } else if (actions.DiscreteActions[0] == 2) - { - moveController.MoveToPointById(actions.DiscreteActions[1]); ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } } #endregion diff --git a/Assets/Scripts/Misc/NavPoint.cs b/Assets/Scripts/Misc/NavPoint.cs index 473d7fa..a70fd61 100755 --- a/Assets/Scripts/Misc/NavPoint.cs +++ b/Assets/Scripts/Misc/NavPoint.cs @@ -17,21 +17,12 @@ public class NavPoint : MonoBehaviour public NavPointType navType = NavPointType.Direction; [HideInInspector] -<<<<<<< HEAD public int PointId = 0; -======= - public int PointId; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 public float DeathAttr = 0; public List EnemiesSeen = new List(); private void Start() { -<<<<<<< HEAD FlagDistance = (GameObject.FindGameObjectWithTag("Flag").transform.position - Position).magnitude; -======= - PointId = GetInstanceID(); - FlagDistance = (GameObject.FindGameObjectWithTag("Flag").transform.position - position).magnitude; ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 } } -- 2.49.0 From 2caf658d190d294de644e3042211c96566daa261 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Thu, 5 May 2022 18:11:57 +0700 Subject: [PATCH 10/17] "update unresolved conflicts" --- Assets/Scripts/Character/CharacterCondition.cs | 1 - Assets/Scripts/Character/MovementController.cs | 10 ---------- Assets/Scripts/Managers/MapManager.cs | 2 -- 3 files changed, 13 deletions(-) diff --git a/Assets/Scripts/Character/CharacterCondition.cs b/Assets/Scripts/Character/CharacterCondition.cs index 8d7d5c5..2124ddb 100755 --- a/Assets/Scripts/Character/CharacterCondition.cs +++ b/Assets/Scripts/Character/CharacterCondition.cs @@ -9,7 +9,6 @@ public class CharacterCondition public event Action OnChangeAmmunitionEvent; private int health; - public int HealthPoints { get diff --git a/Assets/Scripts/Character/MovementController.cs b/Assets/Scripts/Character/MovementController.cs index b1849ce..70341bb 100644 --- a/Assets/Scripts/Character/MovementController.cs +++ b/Assets/Scripts/Character/MovementController.cs @@ -36,16 +36,6 @@ public class MovementController : MonoBehaviour { FlagDistance = (flag.transform.position - gameObject.transform.position).magnitude; } - - public void MoveToPointById(int id) - { - if (!navPoints.ContainsKey(id)) - { - Debug.LogWarning("PIDOR"); - return; - } - goToNextNavPoint(navPoints[id]); - } public void MoveToRandomPoint() { diff --git a/Assets/Scripts/Managers/MapManager.cs b/Assets/Scripts/Managers/MapManager.cs index b28dcba..58e5c77 100755 --- a/Assets/Scripts/Managers/MapManager.cs +++ b/Assets/Scripts/Managers/MapManager.cs @@ -5,8 +5,6 @@ public class MapManager : MonoBehaviour { private static MapManager instance; public static MapManager Instance => instance; - private static List navPoints = new List(); - private static Dictionary iDToNavPoint = new Dictionary(); public static List NavPoints { get => navPoints; private set => navPoints = value; } public static Dictionary IDToNavPoint { get => iDToNavPoint; private set => iDToNavPoint = value; } -- 2.49.0 From b8e8e74ab448c9126b0c22b32ed7b1d25507d97d Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Thu, 5 May 2022 20:45:26 +0700 Subject: [PATCH 11/17] Fixes in Map Manager --- Assets/Scripts/Managers/GameManager.cs | 8 ++++---- Assets/Scripts/Managers/GlobalEventManager.cs | 12 ++++++------ Assets/Scripts/Managers/MapManager.cs | 5 +++-- Assets/Scripts/Statistics/Logger.cs | 14 +++++++------- Assets/Scripts/Statistics/StatisticManager.cs | 4 ++-- 5 files changed, 22 insertions(+), 21 deletions(-) diff --git a/Assets/Scripts/Managers/GameManager.cs b/Assets/Scripts/Managers/GameManager.cs index b6467d8..9299c9b 100755 --- a/Assets/Scripts/Managers/GameManager.cs +++ b/Assets/Scripts/Managers/GameManager.cs @@ -25,8 +25,8 @@ public class GameManager : MonoBehaviour { Academy.Instance.OnEnvironmentReset += ResetScene; - GlobalEventManager.onCaptureFlag += FlagCaptured; - GlobalEventManager.onTimeLeft += TimeOut; + GlobalEventManager.OnCaptureFlag += FlagCaptured; + GlobalEventManager.OnTimeLeft += TimeOut; var agents = GameObject.FindObjectsOfType(); foreach (var item in agents) @@ -138,8 +138,8 @@ public class GameManager : MonoBehaviour private void OnDestroy() { - GlobalEventManager.onCaptureFlag -= FlagCaptured; - GlobalEventManager.onTimeLeft -= TimeOut; + GlobalEventManager.OnCaptureFlag -= FlagCaptured; + GlobalEventManager.OnTimeLeft -= TimeOut; } public static event Action OnResetScene; diff --git a/Assets/Scripts/Managers/GlobalEventManager.cs b/Assets/Scripts/Managers/GlobalEventManager.cs index e0461af..753b7d2 100755 --- a/Assets/Scripts/Managers/GlobalEventManager.cs +++ b/Assets/Scripts/Managers/GlobalEventManager.cs @@ -2,18 +2,18 @@ public class GlobalEventManager { - public static event Action onCaptureFlag; + public static event Action OnCaptureFlag; public static void SendCaptureFlag(Team team) { - onCaptureFlag?.Invoke(team); - onCaptureFlag = null; + OnCaptureFlag?.Invoke(team); + OnCaptureFlag = null; } - public static event Action onTimeLeft; + public static event Action OnTimeLeft; public static void SendTimeout() { - onTimeLeft?.Invoke(); - onTimeLeft = null; + OnTimeLeft?.Invoke(); + OnTimeLeft = null; } } diff --git a/Assets/Scripts/Managers/MapManager.cs b/Assets/Scripts/Managers/MapManager.cs index 58e5c77..14d70af 100755 --- a/Assets/Scripts/Managers/MapManager.cs +++ b/Assets/Scripts/Managers/MapManager.cs @@ -5,8 +5,9 @@ public class MapManager : MonoBehaviour { private static MapManager instance; public static MapManager Instance => instance; - public static List NavPoints { get => navPoints; private set => navPoints = value; } - public static Dictionary IDToNavPoint { get => iDToNavPoint; private set => iDToNavPoint = value; } + [SerializeField] static List _navPoints; + public static List NavPoints { get => _navPoints; private set => _navPoints = value; } + public static Dictionary IDToNavPoint {get; private set; } private void Awake() { diff --git a/Assets/Scripts/Statistics/Logger.cs b/Assets/Scripts/Statistics/Logger.cs index e293d25..df7ce3d 100644 --- a/Assets/Scripts/Statistics/Logger.cs +++ b/Assets/Scripts/Statistics/Logger.cs @@ -3,17 +3,17 @@ using UnityEngine; public class Logger { - private const string directory = "/Logs/"; - private const string baseName = "Log#"; + private const string Directory = "/Logs/"; + private const string BaseName = "Log#"; public static void SaveLog(T objToSerialize) { - string dir = Application.persistentDataPath + directory; - if (!Directory.Exists(dir)) - Directory.CreateDirectory(dir); + var dir = Application.persistentDataPath + Directory; + if (!System.IO.Directory.Exists(dir)) + System.IO.Directory.CreateDirectory(dir); - var logName = baseName + (Directory.GetFiles(dir).Length + 1).ToString(); - string json = JsonUtility.ToJson(objToSerialize); + var logName = BaseName + (System.IO.Directory.GetFiles(dir).Length + 1).ToString(); + var json = JsonUtility.ToJson(objToSerialize); File.WriteAllText(dir + logName, json); } } \ No newline at end of file diff --git a/Assets/Scripts/Statistics/StatisticManager.cs b/Assets/Scripts/Statistics/StatisticManager.cs index 399e7fb..a9c4883 100644 --- a/Assets/Scripts/Statistics/StatisticManager.cs +++ b/Assets/Scripts/Statistics/StatisticManager.cs @@ -19,8 +19,8 @@ public class StatisticManager : MonoBehaviour foreach (var npc in GameObject.FindObjectsOfType()) npc.OnDamageRecieved += RegisterDamage; - GlobalEventManager.onCaptureFlag += RegisterWin; - GlobalEventManager.onTimeLeft += RegisterTimeOut; + GlobalEventManager.OnCaptureFlag += RegisterWin; + GlobalEventManager.OnTimeLeft += RegisterTimeOut; } private void RegisterDamage(int damage, Team team) -- 2.49.0 From 3908334f656b248dc1f2eb4fe5a7918685dd472e Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Sat, 7 May 2022 21:45:28 +0700 Subject: [PATCH 12/17] Finally fixed --- Assets/ML-Agents/Timers/Greatest_map_ever_timers.json.meta | 7 ------- ProjectSettings/ProjectVersion.txt | 2 -- 2 files changed, 9 deletions(-) delete mode 100755 Assets/ML-Agents/Timers/Greatest_map_ever_timers.json.meta delete mode 100755 ProjectSettings/ProjectVersion.txt diff --git a/Assets/ML-Agents/Timers/Greatest_map_ever_timers.json.meta b/Assets/ML-Agents/Timers/Greatest_map_ever_timers.json.meta deleted file mode 100755 index 1400775..0000000 --- a/Assets/ML-Agents/Timers/Greatest_map_ever_timers.json.meta +++ /dev/null @@ -1,7 +0,0 @@ -fileFormatVersion: 2 -guid: 80ef0d75029e25243857877facd14d75 -TextScriptImporter: - externalObjects: {} - userData: - assetBundleName: - assetBundleVariant: diff --git a/ProjectSettings/ProjectVersion.txt b/ProjectSettings/ProjectVersion.txt deleted file mode 100755 index bf8f7eb..0000000 --- a/ProjectSettings/ProjectVersion.txt +++ /dev/null @@ -1,2 +0,0 @@ -m_EditorVersion: 2019.4.36f1 -m_EditorVersionWithRevision: 2019.4.36f1 (660c164b2fc5) -- 2.49.0 From 87568c4077fa09ca9106d5a544aa848f754a6d41 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Sat, 7 May 2022 22:00:19 +0700 Subject: [PATCH 13/17] added scene --- .../Greatest_map_ever/Greatest_map_ever.unity | 212 +----------------- 1 file changed, 6 insertions(+), 206 deletions(-) mode change 100755 => 100644 Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity diff --git a/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity b/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity old mode 100755 new mode 100644 index 19be7a0..1b0de26 --- a/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity +++ b/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity @@ -38,7 +38,7 @@ RenderSettings: m_ReflectionIntensity: 1 m_CustomReflection: {fileID: 0} m_Sun: {fileID: 705507994} - m_IndirectSpecularColor: {r: 0.44657815, g: 0.49641192, b: 0.57481617, a: 1} + m_IndirectSpecularColor: {r: 0.44657898, g: 0.4964133, b: 0.5748178, a: 1} m_UseRadianceAmbientProbe: 0 --- !u!157 &3 LightmapSettings: @@ -259,10 +259,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: -<<<<<<< HEAD navType: 1 -======= ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 PointId: 0 DeathAttr: 0 EnemiesSeen: [] @@ -358,10 +355,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: -<<<<<<< HEAD navType: 1 -======= ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 PointId: 0 DeathAttr: 0 EnemiesSeen: [] @@ -1130,7 +1124,7 @@ Transform: m_PrefabAsset: {fileID: 0} m_GameObject: {fileID: 705507993} m_LocalRotation: {x: 0.40821788, y: -0.23456968, z: 0.10938163, w: 0.8754261} - m_LocalPosition: {x: 81.2, y: 64.1, z: -12.4} + m_LocalPosition: {x: 77.5, y: 55.2, z: -5.9} m_LocalScale: {x: 1, y: 1, z: 1} m_Children: [] m_Father: {fileID: 0} @@ -1261,12 +1255,6 @@ Transform: m_Father: {fileID: 671439045} m_RootOrder: 1 m_LocalEulerAnglesHint: {x: 0, y: 90, z: 0} ---- !u!4 &868386701 stripped -Transform: - m_CorrespondingSourceObject: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - m_PrefabInstance: {fileID: 1809549197} - m_PrefabAsset: {fileID: 0} --- !u!1 &884498019 GameObject: m_ObjectHideFlags: 0 @@ -1300,80 +1288,6 @@ Transform: m_Father: {fileID: 782729761} m_RootOrder: 0 m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} ---- !u!1001 &1061293905 -PrefabInstance: - m_ObjectHideFlags: 0 - serializedVersion: 2 - m_Modification: - m_TransformParent: {fileID: 868386701} - m_Modifications: - - target: {fileID: 2836004473841745626, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_Controller - value: - objectReference: {fileID: 9100000, guid: 95a44e56d04c7d248ba723eda9611c51, type: 2} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_RootOrder - value: 2 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalPosition.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalPosition.y - value: -1 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalPosition.z - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalRotation.w - value: 1 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalRotation.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalRotation.y - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalRotation.z - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalEulerAnglesHint.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalEulerAnglesHint.y - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591197023, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_LocalEulerAnglesHint.z - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 2836004475591695743, guid: f432554f564e69242897607d34218939, - type: 3} - propertyPath: m_Name - value: DragonFucker - objectReference: {fileID: 0} - m_RemovedComponents: [] - m_SourcePrefab: {fileID: 100100000, guid: f432554f564e69242897607d34218939, type: 3} --- !u!1 &1116745543 GameObject: m_ObjectHideFlags: 0 @@ -1405,10 +1319,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: -<<<<<<< HEAD navType: 1 -======= ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 PointId: 0 DeathAttr: 0 EnemiesSeen: [] @@ -1840,7 +1751,7 @@ Transform: m_PrefabAsset: {fileID: 0} m_GameObject: {fileID: 1345085340} m_LocalRotation: {x: -0, y: -0, z: -0, w: 1} - m_LocalPosition: {x: 21.51, y: -5.782543, z: -10.46} + m_LocalPosition: {x: 47.446796, y: -5.782543, z: -25.400002} m_LocalScale: {x: 1.4367, y: 1.4367, z: 1.4367} m_Children: [] m_Father: {fileID: 2060099472} @@ -1858,10 +1769,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: -<<<<<<< HEAD navType: 1 -======= ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 PointId: 0 DeathAttr: 0 EnemiesSeen: [] @@ -1957,10 +1865,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: -<<<<<<< HEAD navType: 1 -======= ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 PointId: 0 DeathAttr: 0 EnemiesSeen: [] @@ -2182,10 +2087,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: -<<<<<<< HEAD navType: 1 -======= ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 PointId: 0 DeathAttr: 0 EnemiesSeen: [] @@ -2236,105 +2138,6 @@ MeshFilter: m_PrefabAsset: {fileID: 0} m_GameObject: {fileID: 1663305221} m_Mesh: {fileID: 10206, guid: 0000000000000000e000000000000000, type: 0} ---- !u!1001 &1809549197 -PrefabInstance: - m_ObjectHideFlags: 0 - serializedVersion: 2 - m_Modification: - m_TransformParent: {fileID: 0} - m_Modifications: - - target: {fileID: -4942972567661207728, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_Enabled - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_RootOrder - value: 6 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalPosition.x - value: 42.23 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalPosition.y - value: 1 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalPosition.z - value: -15.91 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalRotation.w - value: 1 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalRotation.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalRotation.y - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalRotation.z - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalEulerAnglesHint.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalEulerAnglesHint.y - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763689, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalEulerAnglesHint.z - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763699, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_Name - value: Player - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763700, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: playerStance - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763700, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: playerCrouchStance.CameraHeight - value: 0.3 - objectReference: {fileID: 0} - - target: {fileID: 4528203470625763703, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_Enabled - value: 1 - objectReference: {fileID: 0} - - target: {fileID: 4528203471293941515, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalPosition.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 4528203471293941515, guid: 99bb17a23a489624baeaf337f91a4f84, - type: 3} - propertyPath: m_LocalPosition.z - value: -2.417 - objectReference: {fileID: 0} - m_RemovedComponents: [] - m_SourcePrefab: {fileID: 100100000, guid: 99bb17a23a489624baeaf337f91a4f84, type: 3} --- !u!1 &1858987083 GameObject: m_ObjectHideFlags: 0 @@ -2459,9 +2262,9 @@ MonoBehaviour: - {fileID: 140697606} spawnPointsForAttackersTeam: - {fileID: 2004854094} - AIPrefab: {fileID: 2988578997639256874, guid: b016874eb34cc084aa4359f0bbec50e1, + AIPrefab: {fileID: 6171680323948707524, guid: b016874eb34cc084aa4359f0bbec50e1, type: 3} - PlayerPrefab: {fileID: 5245491127989480125, guid: 99bb17a23a489624baeaf337f91a4f84, + PlayerPrefab: {fileID: 5245491127989480125, guid: 80f6c1c85e5daed4c96c70205ed5503d, type: 3} --- !u!114 &1858987091 MonoBehaviour: @@ -2728,10 +2531,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: -<<<<<<< HEAD navType: 1 -======= ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 PointId: 0 DeathAttr: 0 EnemiesSeen: [] @@ -2820,7 +2620,7 @@ Transform: m_PrefabAsset: {fileID: 0} m_GameObject: {fileID: 2060099471} m_LocalRotation: {x: -0, y: -0, z: -0, w: 1} - m_LocalPosition: {x: 27.79, y: 7.782543, z: -15} + m_LocalPosition: {x: 19.553204, y: 7.782543, z: -15} m_LocalScale: {x: 1, y: 1, z: 1} m_Children: - {fileID: 1345085341} -- 2.49.0 From f9a3b06ec37f5b9e18b1c6a89fa972fbb20116a5 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Sat, 7 May 2022 22:02:15 +0700 Subject: [PATCH 14/17] files add --- Assets/Prefabs/Bot.prefab | 237 +++++++++++------- Assets/Prefabs/Player.prefab | 150 ++++++----- .../Scripts/Character/MovementController.cs | 21 +- Assets/Scripts/Character/NPC.cs | 26 +- Assets/Scripts/Managers/GameManager.cs | 1 - Assets/Scripts/Managers/MapManager.cs | 30 +-- .../Statistics/StatisticManager.cs.meta | 2 +- Assets/Settings/Game Settings.asset | 27 -- 8 files changed, 281 insertions(+), 213 deletions(-) diff --git a/Assets/Prefabs/Bot.prefab b/Assets/Prefabs/Bot.prefab index 0af3f18..5b177e0 100755 --- a/Assets/Prefabs/Bot.prefab +++ b/Assets/Prefabs/Bot.prefab @@ -1,6 +1,6 @@ %YAML 1.1 %TAG !u! tag:unity3d.com,2011: ---- !u!1 &2988578997639256874 +--- !u!1 &6171680323407988095 GameObject: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} @@ -8,52 +8,85 @@ GameObject: m_PrefabAsset: {fileID: 0} serializedVersion: 6 m_Component: - - component: {fileID: 2988578997639256870} - - component: {fileID: 2988578997639256869} - - component: {fileID: 2988578997639256868} - - component: {fileID: 2988578997639256875} - - component: {fileID: 5447337162552783061} - - component: {fileID: 7805954453358028498} - - component: {fileID: 2676446634235362783} - - component: {fileID: 8656710265340117963} - - component: {fileID: 778652956973742106} - - component: {fileID: 1208561866453126566} - m_Layer: 0 - m_Name: Bot + - component: {fileID: 6171680323407988094} + m_Layer: 8 + m_Name: FeetTransform m_TagString: Untagged m_Icon: {fileID: 0} m_NavMeshLayer: 0 m_StaticEditorFlags: 0 m_IsActive: 1 ---- !u!4 &2988578997639256870 +--- !u!4 &6171680323407988094 Transform: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323407988095} m_LocalRotation: {x: 0, y: 0, z: 0, w: 1} - m_LocalPosition: {x: 28.310326, y: 13.98, z: 46.45} + m_LocalPosition: {x: 0, y: -1, z: 0} m_LocalScale: {x: 1, y: 1, z: 1} m_Children: [] - m_Father: {fileID: 0} + m_Father: {fileID: 6171680323948707550} m_RootOrder: 0 m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} ---- !u!33 &2988578997639256869 +--- !u!1 &6171680323948707524 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6171680323948707550} + - component: {fileID: 6171680323948707551} + - component: {fileID: 5770895893828047079} + - component: {fileID: 6171680323948707521} + - component: {fileID: 6171680323948707549} + - component: {fileID: 6171680323948707520} + - component: {fileID: 8774702625908438859} + - component: {fileID: 6521536090983603910} + - component: {fileID: 6133354754598649724} + - component: {fileID: 2756943273076691504} + - component: {fileID: 6638271233700792696} + - component: {fileID: 5988018701276467001} + m_Layer: 8 + m_Name: Bot + m_TagString: Defender + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6171680323948707550 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6171680323948707524} + m_LocalRotation: {x: 0, y: 0.7071068, z: 0, w: 0.7071068} + m_LocalPosition: {x: 50.67923, y: 1.16, z: -22.57} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6171680323407988094} + m_Father: {fileID: 0} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 90, z: 0} +--- !u!33 &6171680323948707551 MeshFilter: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323948707524} m_Mesh: {fileID: 10208, guid: 0000000000000000e000000000000000, type: 0} ---- !u!23 &2988578997639256868 +--- !u!23 &5770895893828047079 MeshRenderer: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323948707524} m_Enabled: 1 m_CastShadows: 1 m_ReceiveShadows: 1 @@ -86,94 +119,58 @@ MeshRenderer: m_SortingLayerID: 0 m_SortingLayer: 0 m_SortingOrder: 0 ---- !u!136 &2988578997639256875 +--- !u!136 &6171680323948707521 CapsuleCollider: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323948707524} m_Material: {fileID: 0} m_IsTrigger: 0 - m_Enabled: 1 - m_Radius: 0.5 + m_Enabled: 0 + m_Radius: 0.3 m_Height: 2 m_Direction: 1 m_Center: {x: 0, y: 0, z: 0} ---- !u!114 &5447337162552783061 -MonoBehaviour: +--- !u!136 &6171680323948707549 +CapsuleCollider: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} - m_Enabled: 1 - m_EditorHideFlags: 0 - m_Script: {fileID: 11500000, guid: a6f2a081cfc8c4b4bb6864331109d147, type: 3} - m_Name: - m_EditorClassIdentifier: - agentParameters: - maxStep: 0 - hasUpgradedFromAgentParameters: 1 - MaxStep: 100 ---- !u!114 &7805954453358028498 -MonoBehaviour: + m_GameObject: {fileID: 6171680323948707524} + m_Material: {fileID: 0} + m_IsTrigger: 0 + m_Enabled: 0 + m_Radius: 0.3 + m_Height: 1.3 + m_Direction: 1 + m_Center: {x: 0, y: -0.35, z: 0} +--- !u!136 &6171680323948707520 +CapsuleCollider: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323948707524} + m_Material: {fileID: 0} + m_IsTrigger: 0 m_Enabled: 1 - m_EditorHideFlags: 0 - m_Script: {fileID: 11500000, guid: 3a5c9d521e5ef4759a8246a07d52221e, type: 3} - m_Name: - m_EditorClassIdentifier: - DecisionPeriod: 1 - TakeActionsBetweenDecisions: 1 ---- !u!114 &2676446634235362783 -MonoBehaviour: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} - m_Enabled: 1 - m_EditorHideFlags: 0 - m_Script: {fileID: 11500000, guid: 5d1c4e0b1822b495aa52bc52839ecb30, type: 3} - m_Name: - m_EditorClassIdentifier: - m_BrainParameters: -<<<<<<< HEAD - VectorObservationSize: 9 -======= - VectorObservationSize: 4 ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 - NumStackedVectorObservations: 1 - m_ActionSpec: - m_NumContinuousActions: 0 - BranchSizes: 0200000002000000 - VectorActionSize: 0200000002000000 - VectorActionDescriptions: [] - VectorActionSpaceType: 0 - hasUpgradedBrainParametersWithActionSpec: 1 - m_Model: {fileID: 0} - m_InferenceDevice: 0 - m_BehaviorType: 0 - m_BehaviorName: npc - TeamId: 0 - m_UseChildSensors: 1 - m_UseChildActuators: 1 - m_ObservableAttributeHandling: 0 ---- !u!195 &8656710265340117963 + m_Radius: 0.3 + m_Height: 0.8 + m_Direction: 1 + m_Center: {x: 0, y: -0.6, z: 0} +--- !u!195 &8774702625908438859 NavMeshAgent: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323948707524} m_Enabled: 1 m_AgentTypeID: 0 - m_Radius: 0.5 + m_Radius: 0.5000001 m_Speed: 3.5 m_Acceleration: 8 avoidancePriority: 50 @@ -186,32 +183,92 @@ NavMeshAgent: m_BaseOffset: 1 m_WalkableMask: 4294967295 m_ObstacleAvoidanceType: 4 ---- !u!114 &778652956973742106 +--- !u!114 &6521536090983603910 MonoBehaviour: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323948707524} m_Enabled: 1 m_EditorHideFlags: 0 m_Script: {fileID: 11500000, guid: d3ebcf807a37f344998fd648dfc9376d, type: 3} m_Name: m_EditorClassIdentifier: - navMeshAgent: {fileID: 8656710265340117963} + navMeshAgent: {fileID: 8774702625908438859} flag: {fileID: 6818223691859422291, guid: 1685c1d9ce4ab174f95c646b1826010b, type: 3} ---- !u!114 &1208561866453126566 +--- !u!114 &6133354754598649724 MonoBehaviour: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 2988578997639256874} + m_GameObject: {fileID: 6171680323948707524} m_Enabled: 1 m_EditorHideFlags: 0 m_Script: {fileID: 11500000, guid: dd8012d5925524537b27131fef517017, type: 3} m_Name: m_EditorClassIdentifier: m_SensorName: BufferSensor - m_ObservableSize: 5 - m_MaxNumObservables: 5 + m_ObservableSize: 0 + m_MaxNumObservables: 0 +--- !u!114 &2756943273076691504 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6171680323948707524} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 5d1c4e0b1822b495aa52bc52839ecb30, type: 3} + m_Name: + m_EditorClassIdentifier: + m_BrainParameters: + VectorObservationSize: 1 + NumStackedVectorObservations: 1 + m_ActionSpec: + m_NumContinuousActions: 0 + BranchSizes: 01000000 + VectorActionSize: 01000000 + VectorActionDescriptions: [] + VectorActionSpaceType: 0 + hasUpgradedBrainParametersWithActionSpec: 1 + m_Model: {fileID: 0} + m_InferenceDevice: 0 + m_BehaviorType: 0 + m_BehaviorName: My Behavior + TeamId: 0 + m_UseChildSensors: 1 + m_UseChildActuators: 1 + m_ObservableAttributeHandling: 0 +--- !u!114 &6638271233700792696 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6171680323948707524} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: a6f2a081cfc8c4b4bb6864331109d147, type: 3} + m_Name: + m_EditorClassIdentifier: + agentParameters: + maxStep: 0 + hasUpgradedFromAgentParameters: 1 + MaxStep: 100 +--- !u!114 &5988018701276467001 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6171680323948707524} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: d3ebcf807a37f344998fd648dfc9376d, type: 3} + m_Name: + m_EditorClassIdentifier: + navMeshAgent: {fileID: 8774702625908438859} + flag: {fileID: 6818223691859422291, guid: 1685c1d9ce4ab174f95c646b1826010b, type: 3} diff --git a/Assets/Prefabs/Player.prefab b/Assets/Prefabs/Player.prefab index c798d61..31319fc 100755 --- a/Assets/Prefabs/Player.prefab +++ b/Assets/Prefabs/Player.prefab @@ -14,9 +14,12 @@ GameObject: - component: {fileID: 5245491127989480120} - component: {fileID: 5245491127989480100} - component: {fileID: 5245491127989480121} - - component: {fileID: 5245491127989480123} - - component: {fileID: 5245491127989480122} - - component: {fileID: 4890899368932544690} + - component: {fileID: 7254047075221496626} + - component: {fileID: 5182704636738128575} + - component: {fileID: 5220658550450318085} + - component: {fileID: 4264677542023120457} + - component: {fileID: 4715950961444674817} + - component: {fileID: 5653209015090846528} m_Layer: 8 m_Name: Player m_TagString: Defender @@ -129,25 +132,29 @@ CapsuleCollider: m_Height: 0.8 m_Direction: 1 m_Center: {x: 0, y: -0.6, z: 0} ---- !u!143 &5245491127989480123 -CharacterController: +--- !u!195 &7254047075221496626 +NavMeshAgent: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} m_GameObject: {fileID: 5245491127989480125} - m_Material: {fileID: 0} - m_IsTrigger: 0 m_Enabled: 1 - serializedVersion: 2 + m_AgentTypeID: 0 + m_Radius: 0.5000001 + m_Speed: 3.5 + m_Acceleration: 8 + avoidancePriority: 50 + m_AngularSpeed: 120 + m_StoppingDistance: 0 + m_AutoTraverseOffMeshLink: 1 + m_AutoBraking: 1 + m_AutoRepath: 1 m_Height: 2 - m_Radius: 0.5 - m_SlopeLimit: 45 - m_StepOffset: 0.3 - m_SkinWidth: 0.08 - m_MinMoveDistance: 0.001 - m_Center: {x: 0, y: 0, z: 0} ---- !u!114 &5245491127989480122 + m_BaseOffset: 1 + m_WalkableMask: 4294967295 + m_ObstacleAvoidanceType: 4 +--- !u!114 &5182704636738128575 MonoBehaviour: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} @@ -156,52 +163,12 @@ MonoBehaviour: m_GameObject: {fileID: 5245491127989480125} m_Enabled: 1 m_EditorHideFlags: 0 - m_Script: {fileID: 11500000, guid: 9826297ef4d853741b2af768441ec7f7, type: 3} + m_Script: {fileID: 11500000, guid: d3ebcf807a37f344998fd648dfc9376d, type: 3} m_Name: m_EditorClassIdentifier: - input_View: {x: 0, y: 0} - cameraHolder: {fileID: 5245491129196666053} - feetTransform: {fileID: 5245491129603592455} - playerSettings: - ViewXSensetivity: 20 - ViewYSensetivity: 20 - ViewXInverted: 0 - ViewYInverted: 0 - SprintingHold: 0 - MovementSmoothing: 0 - RunningForwardSpeed: 10 - RunningStrafeSpeed: 6 - WalkingForwardSpeed: 4 - WalkingBackwardSpeed: 2 - WalkingStrafeSpeed: 3 - JumpingHeight: 6 - JumpingFalloff: 1 - FallingSmoothing: 0 - SpeedEffector: 1 - CrouchSpeedEffector: 0 - ProneSpeedEffector: 0 - FallingSpeedEffector: 0 - ViewClampYMin: -70 - ViewClampYMax: 80 - playerMask: - serializedVersion: 2 - m_Bits: 55 - gravityAmount: 0.05 - gravityMin: -3 - jumpingForce: {x: 0, y: 0, z: 0} - playerStance: 0 - playerStanceSmoothing: 0.2 - playerStandStance: - CameraHeight: 0.7 - StanceCollider: {fileID: 5245491127989480120} - playerCrouchStance: - CameraHeight: 0.3 - StanceCollider: {fileID: 5245491127989480100} - playerProneStance: - CameraHeight: -0.58 - StanceCollider: {fileID: 5245491127989480121} - currentWeapon: {fileID: 8510909888689775086} ---- !u!114 &4890899368932544690 + navMeshAgent: {fileID: 0} + flag: {fileID: 0} +--- !u!114 &5220658550450318085 MonoBehaviour: m_ObjectHideFlags: 0 m_CorrespondingSourceObject: {fileID: 0} @@ -210,9 +177,72 @@ MonoBehaviour: m_GameObject: {fileID: 5245491127989480125} m_Enabled: 1 m_EditorHideFlags: 0 - m_Script: {fileID: 11500000, guid: a8c9a8e604d395c4ab9d03d28adc4982, type: 3} + m_Script: {fileID: 11500000, guid: dd8012d5925524537b27131fef517017, type: 3} m_Name: m_EditorClassIdentifier: + m_SensorName: BufferSensor + m_ObservableSize: 0 + m_MaxNumObservables: 0 +--- !u!114 &4264677542023120457 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 5245491127989480125} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 5d1c4e0b1822b495aa52bc52839ecb30, type: 3} + m_Name: + m_EditorClassIdentifier: + m_BrainParameters: + VectorObservationSize: 1 + NumStackedVectorObservations: 1 + m_ActionSpec: + m_NumContinuousActions: 0 + BranchSizes: 01000000 + VectorActionSize: 01000000 + VectorActionDescriptions: [] + VectorActionSpaceType: 0 + hasUpgradedBrainParametersWithActionSpec: 1 + m_Model: {fileID: 0} + m_InferenceDevice: 0 + m_BehaviorType: 0 + m_BehaviorName: My Behavior + TeamId: 0 + m_UseChildSensors: 1 + m_UseChildActuators: 1 + m_ObservableAttributeHandling: 0 +--- !u!114 &4715950961444674817 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 5245491127989480125} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: a6f2a081cfc8c4b4bb6864331109d147, type: 3} + m_Name: + m_EditorClassIdentifier: + agentParameters: + maxStep: 0 + hasUpgradedFromAgentParameters: 1 + MaxStep: 0 +--- !u!114 &5653209015090846528 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 5245491127989480125} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: d3ebcf807a37f344998fd648dfc9376d, type: 3} + m_Name: + m_EditorClassIdentifier: + navMeshAgent: {fileID: 7254047075221496626} + flag: {fileID: 0} --- !u!1 &5245491128202443531 GameObject: m_ObjectHideFlags: 0 diff --git a/Assets/Scripts/Character/MovementController.cs b/Assets/Scripts/Character/MovementController.cs index 70341bb..98df2aa 100644 --- a/Assets/Scripts/Character/MovementController.cs +++ b/Assets/Scripts/Character/MovementController.cs @@ -9,27 +9,28 @@ public class MovementController : MonoBehaviour public int PointStartID { get; set; } public int PointEndID { get; private set; } public float FlagDistance { get; private set; } - private const float updateFlagPositionDelay = 5; - private const float updateReachedDestinationDelay = 5; + private const float UpdateFlagPositionDelay = 5; + private const float UpdateReachedDestinationDelay = 5; [SerializeField] private NavMeshAgent navMeshAgent; [SerializeField] private GameObject flag; public float DistanceToGo { get; private set; } public float RemainingDistance => navMeshAgent.remainingDistance; - private Dictionary idNavPointDict; + private Dictionary _idNavPointDict; private void Awake() { navMeshAgent.speed = SettingsReader.Instance.GetSettings.MovementSpeed; - idNavPointDict = MapManager.IDToNavPoint; - InvokeRepeating(nameof(UpdateFlagPosition), 0, updateFlagPositionDelay); - InvokeRepeating(nameof(ReachedDestination), 0, updateReachedDestinationDelay); + _idNavPointDict = MapManager.Instance.IDToNavPoint; + InvokeRepeating(nameof(UpdateFlagPosition), 0, UpdateFlagPositionDelay); + InvokeRepeating(nameof(ReachedDestination), 0, UpdateReachedDestinationDelay); } private void OnDestroy() { CancelInvoke(nameof(UpdateFlagPosition)); + CancelInvoke(nameof(ReachedDestination)); } private void UpdateFlagPosition() @@ -39,14 +40,14 @@ public class MovementController : MonoBehaviour public void MoveToRandomPoint() { - GoToNextNavPoint(MapManager.NavPoints[Random.Range(0, MapManager.NavPoints.Count)]); + GoToNextNavPoint(MapManager.Instance.NavPoints[Random.Range(0, MapManager.Instance.NavPoints.Count)]); } public List GetPointsCandidate() { - return MapManager.NavPoints + return MapManager.Instance.NavPoints .Where(point => - (idNavPointDict[PointStartID].Position - point.Position).magnitude < SettingsReader.Instance.GetSettings.MovementDistance) + (_idNavPointDict[PointStartID].Position - point.Position).magnitude < SettingsReader.Instance.GetSettings.MovementDistance) .ToList(); } @@ -62,7 +63,7 @@ public class MovementController : MonoBehaviour public void ReturnToStartPoint() { if (navMeshAgent.isStopped == true) navMeshAgent.isStopped = false; - navMeshAgent.SetDestination(idNavPointDict[PointStartID].Position); + navMeshAgent.SetDestination(_idNavPointDict[PointStartID].Position); PointEndID = PointStartID; PointStartID = -1; } diff --git a/Assets/Scripts/Character/NPC.cs b/Assets/Scripts/Character/NPC.cs index d47e377..36b76ec 100644 --- a/Assets/Scripts/Character/NPC.cs +++ b/Assets/Scripts/Character/NPC.cs @@ -5,7 +5,7 @@ using Unity.MLAgents.Actuators; using Unity.MLAgents.Sensors; using UnityEngine; -[RequireComponent(typeof(MovementController),typeof(BufferSensor))] +[RequireComponent(typeof(MovementController),typeof(BufferSensorComponent))] public class NPC : Agent, ICharacter { [HideInInspector] @@ -47,12 +47,12 @@ public class NPC : Agent, ICharacter moveController = gameObject.GetComponent(); bufferSensor = gameObject.GetComponent(); - + flagZone = GameObject.FindObjectOfType(); - if (flagZone == null) - Debug.LogError("Flag Is Not Setted"); - - navPointIdDict = MapManager.IDToNavPoint; + if (flagZone is null) + Debug.LogError("Flag Is Not Set"); + + navPointIdDict = MapManager.Instance.IDToNavPoint; if (navPointIdDict is null) Debug.LogError("Cant Find Nav Point Dictionary"); } @@ -64,12 +64,18 @@ public class NPC : Agent, ICharacter public override void OnEpisodeBegin() { + if (navPointIdDict is null) + Debug.LogError("Cant Find Nav Point Dictionary"); + NpcState = DirectState; flagZone = GameObject.FindObjectOfType(); } public override void CollectObservations(VectorSensor sensor) { + navPointIdDict = MapManager.Instance.IDToNavPoint; + if (navPointIdDict is null) + Debug.LogError("Cant Find Nav Point Dictionary"); var candidates = moveController.GetPointsCandidate(); //common sensors @@ -93,13 +99,14 @@ public class NPC : Agent, ICharacter //point sensors foreach (var point in candidates) { + var position = transform.position; bufferSensor.AppendObservation(new float[] { point.DeathAttr, (int)point.navType, //4 flagEnemyDistance - GameManager.IsCloserToFlagFromNextNavPoint(point, transform.position).ToInt(), + GameManager.IsCloserToFlagFromNextNavPoint(point, position).ToInt(), //5 EnemyVsNavPointDistance - GameManager.IsCloserToEnemyThanToNextNavPoint(point,transform.position, AgentCharacter.Team.GetOppositeTeam()).ToInt(), + GameManager.IsCloserToEnemyThanToNextNavPoint(point, position, AgentCharacter.Team.GetOppositeTeam()).ToInt(), //6 Have been seen by enemy in this point GameManager.IsHaveSeenByEnemy(AgentCharacter.Team.GetOppositeTeam(), point.Position).ToInt() @@ -165,11 +172,12 @@ public class NPC : Agent, ICharacter } public event Action OnDamageRecieved; - public void GetDamage(float damage) + public void GetDamage(int damage) { AgentCharacter.LastTimeHit = TimeManager.Instance.CurrentTime; Condition.GiveHealth(-Mathf.RoundToInt(damage * (1 - Condition.ArmourPoints * 0.5f))); Condition.GiveArmour(-Mathf.RoundToInt(Mathf.Sqrt(damage) * 5)); + OnDamageRecieved?.Invoke(damage, AgentCharacter.Team); if (Condition.HealthPoints < 0) { diff --git a/Assets/Scripts/Managers/GameManager.cs b/Assets/Scripts/Managers/GameManager.cs index 9299c9b..431fd81 100755 --- a/Assets/Scripts/Managers/GameManager.cs +++ b/Assets/Scripts/Managers/GameManager.cs @@ -24,7 +24,6 @@ public class GameManager : MonoBehaviour private void Start() { Academy.Instance.OnEnvironmentReset += ResetScene; - GlobalEventManager.OnCaptureFlag += FlagCaptured; GlobalEventManager.OnTimeLeft += TimeOut; diff --git a/Assets/Scripts/Managers/MapManager.cs b/Assets/Scripts/Managers/MapManager.cs index 14d70af..97a0c75 100755 --- a/Assets/Scripts/Managers/MapManager.cs +++ b/Assets/Scripts/Managers/MapManager.cs @@ -3,36 +3,36 @@ using UnityEngine; public class MapManager : MonoBehaviour { - private static MapManager instance; - public static MapManager Instance => instance; - [SerializeField] static List _navPoints; - public static List NavPoints { get => _navPoints; private set => _navPoints = value; } - public static Dictionary IDToNavPoint {get; private set; } + private static MapManager _instance; + public static MapManager Instance => _instance; + [SerializeField] private List _navPoints; + public List NavPoints { get => _navPoints; private set => _navPoints = value; } + public Dictionary IDToNavPoint {get; private set; } private void Awake() { - if (instance is null) - instance = this; + if (_instance is null) + _instance = this; else { Destroy(gameObject); Debug.LogError("Only 1 Instance"); } - } - - private void Start() - { + + NavPoints = new List(); var navPointSet = GameObject.Find("NavPoint Set"); var count = navPointSet.transform.childCount; - for (int i=0; i < count; i++) + for (var i=0; i < count; i++) NavPoints.Add(navPointSet.transform.GetChild(i) .gameObject.GetComponent()); - + print(NavPoints.Count); NavPointSetToID(); } + private void NavPointSetToID() { + IDToNavPoint = new Dictionary(); int i = 0; foreach (var navPoint in NavPoints) { @@ -45,8 +45,8 @@ public class MapManager : MonoBehaviour public static void AddDeathAttributeToPoints(int startPoint, int endPoint, float allDistance, float remainingDistance) { - var startNavPoint = IDToNavPoint[startPoint]; - var endNavPoint = IDToNavPoint[endPoint]; + var startNavPoint = _instance.IDToNavPoint[startPoint]; + var endNavPoint = _instance.IDToNavPoint[endPoint]; float coef; try { diff --git a/Assets/Scripts/Statistics/StatisticManager.cs.meta b/Assets/Scripts/Statistics/StatisticManager.cs.meta index 3a27c34..b8be93a 100644 --- a/Assets/Scripts/Statistics/StatisticManager.cs.meta +++ b/Assets/Scripts/Statistics/StatisticManager.cs.meta @@ -4,7 +4,7 @@ MonoImporter: externalObjects: {} serializedVersion: 2 defaultReferences: [] - executionOrder: 300 + executionOrder: 0 icon: {instanceID: 0} userData: assetBundleName: diff --git a/Assets/Settings/Game Settings.asset b/Assets/Settings/Game Settings.asset index ea0fad7..fc07488 100755 --- a/Assets/Settings/Game Settings.asset +++ b/Assets/Settings/Game Settings.asset @@ -12,7 +12,6 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: e2c47233b9062c84482336b145c6891b, type: 3} m_Name: Game Settings m_EditorClassIdentifier: -<<<<<<< HEAD IsTesting: 0 TimeToWin: 15 TimeOut: 3600 @@ -39,29 +38,3 @@ MonoBehaviour: DoDamageChanceInRunning: 30 DoDamageChanceInCover: 25 CrouchingCoefficient: 1.4 -======= - isTesting: 1 - timeToWin: 5 - timeOut: 1600 - movementDistance: 50 - movementSpeed: 3 - defTeamAI: 0 - atcTeamAI: 0 - numOfDefenders: 1 - numOfAttackers: 1 - hasHumanDefender: 1 - hasHumanAttacker: 0 - healthPickupAmount: 50 - armourPickupAmount: 50 - ammunitionPickupAmount: 120 - pickupsAmount: 2 - maxHealth: 100 - maxArmour: 100 - maxAmmo: 31337 - GetHitChanceInDirectPoint: 0 - GetHitChanceInRunning: 0 - GetHitChanceInCover: 0 - DoDamageChanceInDirectPoint: 0 - DoDamageChanceInRunning: 0 - DoDamageChanceInCover: 0 ->>>>>>> 351fa8ee12c499a0374be71a192a5f288611f467 -- 2.49.0 From 8cce356b8e0edda609121fa376809d2e2628ed9d Mon Sep 17 00:00:00 2001 From: Andrey Gumirov Date: Sat, 7 May 2022 22:56:07 +0700 Subject: [PATCH 15/17] MVP work with ml-agents python library --- Assets/Prefabs/Bot.prefab | 27 +++++++-- Assets/Prefabs/Player.prefab.meta | 2 +- Assets/Scripts/Character/NPC.cs | 17 +++++- Assets/Settings/Game Settings.asset | 0 Packages/manifest.json | 7 +++ Packages/packages-lock.json | 79 +++++++++++++++++++++++++++ ProjectSettings/ProjectSettings.asset | 6 +- 7 files changed, 129 insertions(+), 9 deletions(-) mode change 100755 => 100644 Assets/Prefabs/Bot.prefab mode change 100755 => 100644 Assets/Prefabs/Player.prefab.meta mode change 100755 => 100644 Assets/Settings/Game Settings.asset diff --git a/Assets/Prefabs/Bot.prefab b/Assets/Prefabs/Bot.prefab old mode 100755 new mode 100644 index 5b177e0..add7a37 --- a/Assets/Prefabs/Bot.prefab +++ b/Assets/Prefabs/Bot.prefab @@ -50,6 +50,7 @@ GameObject: - component: {fileID: 2756943273076691504} - component: {fileID: 6638271233700792696} - component: {fileID: 5988018701276467001} + - component: {fileID: 1547882613481475944} m_Layer: 8 m_Name: Bot m_TagString: Defender @@ -210,8 +211,8 @@ MonoBehaviour: m_Name: m_EditorClassIdentifier: m_SensorName: BufferSensor - m_ObservableSize: 0 - m_MaxNumObservables: 0 + m_ObservableSize: 5 + m_MaxNumObservables: 10 --- !u!114 &2756943273076691504 MonoBehaviour: m_ObjectHideFlags: 0 @@ -225,19 +226,19 @@ MonoBehaviour: m_Name: m_EditorClassIdentifier: m_BrainParameters: - VectorObservationSize: 1 + VectorObservationSize: 14 NumStackedVectorObservations: 1 m_ActionSpec: m_NumContinuousActions: 0 - BranchSizes: 01000000 - VectorActionSize: 01000000 + BranchSizes: 010000000100000001000000 + VectorActionSize: 010000000100000001000000 VectorActionDescriptions: [] VectorActionSpaceType: 0 hasUpgradedBrainParametersWithActionSpec: 1 m_Model: {fileID: 0} m_InferenceDevice: 0 m_BehaviorType: 0 - m_BehaviorName: My Behavior + m_BehaviorName: npc TeamId: 0 m_UseChildSensors: 1 m_UseChildActuators: 1 @@ -272,3 +273,17 @@ MonoBehaviour: m_EditorClassIdentifier: navMeshAgent: {fileID: 8774702625908438859} flag: {fileID: 6818223691859422291, guid: 1685c1d9ce4ab174f95c646b1826010b, type: 3} +--- !u!114 &1547882613481475944 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6171680323948707524} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 3a5c9d521e5ef4759a8246a07d52221e, type: 3} + m_Name: + m_EditorClassIdentifier: + DecisionPeriod: 5 + TakeActionsBetweenDecisions: 0 diff --git a/Assets/Prefabs/Player.prefab.meta b/Assets/Prefabs/Player.prefab.meta old mode 100755 new mode 100644 index 1fa6659..43cb0ce --- a/Assets/Prefabs/Player.prefab.meta +++ b/Assets/Prefabs/Player.prefab.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: 80f6c1c85e5daed4c96c70205ed5503d +guid: a7480b77908b042d8adcdd84e8c2c15e PrefabImporter: externalObjects: {} userData: diff --git a/Assets/Scripts/Character/NPC.cs b/Assets/Scripts/Character/NPC.cs index 36b76ec..7af2ae9 100644 --- a/Assets/Scripts/Character/NPC.cs +++ b/Assets/Scripts/Character/NPC.cs @@ -73,6 +73,7 @@ public class NPC : Agent, ICharacter public override void CollectObservations(VectorSensor sensor) { + // Debug.Log("Collect observations called!"); navPointIdDict = MapManager.Instance.IDToNavPoint; if (navPointIdDict is null) Debug.LogError("Cant Find Nav Point Dictionary"); @@ -88,6 +89,7 @@ public class NPC : Agent, ICharacter sensor.AddObservation(candidates.Count); sensor.AddObservation(moveController.PointStartID); sensor.AddObservation(moveController.PointEndID); + // Debug.Log("Done common!"); //state sensors sensor.AddObservation((int)NpcState.State); sensor.AddObservation((int)NpcBodyState.State); @@ -95,6 +97,7 @@ public class NPC : Agent, ICharacter sensor.AddObservation(navPointIdDict[moveController.PointStartID].DeathAttr); sensor.AddObservation(navPointIdDict[moveController.PointEndID].DeathAttr); sensor.AddObservation(moveController.FlagDistance); + // Debug.Log("Done state sensors!"); //point sensors foreach (var point in candidates) @@ -112,11 +115,14 @@ public class NPC : Agent, ICharacter point.Position).ToInt() }); } + // Debug.Log("Done collect observations!"); } public override void OnActionReceived(ActionBuffers actions) { + // Debug.Log("Actions recieved!"); var result = actions.DiscreteActions; + // Debug.Log(result[0] + " " + result[1]); if (result[0] == 0) { if (navPointIdDict[moveController.PointStartID].navType != NavPointType.Cover) @@ -132,18 +138,26 @@ public class NPC : Agent, ICharacter default: throw new ArgumentException("Undefined Action recieved"); } } + // Debug.Log(result[0] == 1); if (result[0] == 1) { + // Debug.Log("BEFORE SOme shitty if >:("); if (navPointIdDict[moveController.PointStartID].navType != NavPointType.Direction) + { + // Debug.Log("SOme shitty if >:("); return; + } + // Debug.Log("FUCK"); + switch (result[1]) { case 0: moveController.GoToNextNavPoint(navPointIdDict[result[2]]); - NpcState = RunningState; break; + NpcState = RunningState; Debug.Log("Go to point " + result[2]);break; case 1: NpcState = DirectState; break; default: throw new ArgumentException("Undefined Action recieved"); } } + if (result[0] == 2) { if (moveController.PointStartID == moveController.PointEndID && moveController.PointEndID != -1) @@ -155,6 +169,7 @@ public class NPC : Agent, ICharacter default: throw new ArgumentException("Undefined Action recieved"); } } + // Debug.Log("Actions processed!"); } #endregion diff --git a/Assets/Settings/Game Settings.asset b/Assets/Settings/Game Settings.asset old mode 100755 new mode 100644 diff --git a/Packages/manifest.json b/Packages/manifest.json index a5b072e..e756f82 100755 --- a/Packages/manifest.json +++ b/Packages/manifest.json @@ -1,17 +1,24 @@ { "dependencies": { + "com.unity.2d.sprite": "1.0.0", + "com.unity.2d.tilemap": "1.0.0", + "com.unity.ads": "3.7.5", + "com.unity.analytics": "3.6.12", "com.unity.collab-proxy": "1.14.12", "com.unity.ide.rider": "1.2.1", "com.unity.ide.visualstudio": "2.0.14", "com.unity.ide.vscode": "1.2.4", "com.unity.inputsystem": "1.3.0", "com.unity.ml-agents": "2.0.1", + "com.unity.multiplayer-hlapi": "1.0.8", "com.unity.probuilder": "4.5.2", + "com.unity.purchasing": "4.1.2", "com.unity.test-framework": "1.1.30", "com.unity.textmeshpro": "2.1.6", "com.unity.timeline": "1.2.18", "com.unity.toolchain.win-x86_64-linux-x86_64": "2.0.0", "com.unity.ugui": "1.0.0", + "com.unity.xr.legacyinputhelpers": "2.1.8", "com.unity.modules.ai": "1.0.0", "com.unity.modules.androidjni": "1.0.0", "com.unity.modules.animation": "1.0.0", diff --git a/Packages/packages-lock.json b/Packages/packages-lock.json index bb10cfc..0bac4b2 100755 --- a/Packages/packages-lock.json +++ b/Packages/packages-lock.json @@ -1,5 +1,35 @@ { "dependencies": { + "com.unity.2d.sprite": { + "version": "1.0.0", + "depth": 0, + "source": "builtin", + "dependencies": {} + }, + "com.unity.2d.tilemap": { + "version": "1.0.0", + "depth": 0, + "source": "builtin", + "dependencies": {} + }, + "com.unity.ads": { + "version": "3.7.5", + "depth": 0, + "source": "registry", + "dependencies": { + "com.unity.ugui": "1.0.0" + }, + "url": "https://packages.unity.com" + }, + "com.unity.analytics": { + "version": "3.6.12", + "depth": 0, + "source": "registry", + "dependencies": { + "com.unity.ugui": "1.0.0" + }, + "url": "https://packages.unity.com" + }, "com.unity.barracuda": { "version": "2.0.0", "depth": 1, @@ -86,6 +116,15 @@ }, "url": "https://packages.unity.com" }, + "com.unity.multiplayer-hlapi": { + "version": "1.0.8", + "depth": 0, + "source": "registry", + "dependencies": { + "nuget.mono-cecil": "0.1.6-preview" + }, + "url": "https://packages.unity.com" + }, "com.unity.probuilder": { "version": "4.5.2", "depth": 0, @@ -95,6 +134,29 @@ }, "url": "https://packages.unity.com" }, + "com.unity.purchasing": { + "version": "4.1.2", + "depth": 0, + "source": "registry", + "dependencies": { + "com.unity.ugui": "1.0.0", + "com.unity.modules.unityanalytics": "1.0.0", + "com.unity.modules.unitywebrequest": "1.0.0", + "com.unity.modules.jsonserialize": "1.0.0", + "com.unity.modules.androidjni": "1.0.0", + "com.unity.services.core": "1.0.1" + }, + "url": "https://packages.unity.com" + }, + "com.unity.services.core": { + "version": "1.0.1", + "depth": 1, + "source": "registry", + "dependencies": { + "com.unity.modules.unitywebrequest": "1.0.0" + }, + "url": "https://packages.unity.com" + }, "com.unity.settings-manager": { "version": "1.0.3", "depth": 1, @@ -169,6 +231,23 @@ "com.unity.modules.imgui": "1.0.0" } }, + "com.unity.xr.legacyinputhelpers": { + "version": "2.1.8", + "depth": 0, + "source": "registry", + "dependencies": { + "com.unity.modules.vr": "1.0.0", + "com.unity.modules.xr": "1.0.0" + }, + "url": "https://packages.unity.com" + }, + "nuget.mono-cecil": { + "version": "0.1.6-preview", + "depth": 1, + "source": "registry", + "dependencies": {}, + "url": "https://packages.unity.com" + }, "com.unity.modules.ai": { "version": "1.0.0", "depth": 0, diff --git a/ProjectSettings/ProjectSettings.asset b/ProjectSettings/ProjectSettings.asset index dbf6f6c..b6bb648 100644 --- a/ProjectSettings/ProjectSettings.asset +++ b/ProjectSettings/ProjectSettings.asset @@ -614,7 +614,11 @@ PlayerSettings: webGLLinkerTarget: 1 webGLThreadsSupport: 0 webGLWasmStreaming: 0 - scriptingDefineSymbols: {} + scriptingDefineSymbols: + 1: CROSS_PLATFORM_INPUT + 4: CROSS_PLATFORM_INPUT;MOBILE_INPUT + 7: CROSS_PLATFORM_INPUT;MOBILE_INPUT + 14: MOBILE_INPUT platformArchitecture: {} scriptingBackend: {} il2cppCompilerConfiguration: {} -- 2.49.0 From fa6b5f602ee65d16f94119d8ed74599852c988b4 Mon Sep 17 00:00:00 2001 From: Andrey Gumirov Date: Sat, 7 May 2022 22:58:25 +0700 Subject: [PATCH 16/17] Added test NB --- test-ml-agents.ipynb | 20623 +++++++++++++++++++++++++++++++++++++++++ 1 file changed, 20623 insertions(+) create mode 100644 test-ml-agents.ipynb diff --git a/test-ml-agents.ipynb b/test-ml-agents.ipynb new file mode 100644 index 0000000..1e2d31e --- /dev/null +++ b/test-ml-agents.ipynb @@ -0,0 +1,20623 @@ +{ + "cells": [ + { + "cell_type": "code", + "execution_count": 34, + "id": "8104e2db-b1a8-40a0-a238-5d9149fd74b0", + "metadata": {}, + "outputs": [], + "source": [ + "from mlagents_envs.environment import UnityEnvironment\n", + "import mlagents_envs\n", + "import numpy as np" + ] + }, + { + "cell_type": "code", + "execution_count": 74, + "id": "6f477382-acc9-4aec-907a-7f58caf955ed", + "metadata": {}, + "outputs": [], + "source": [ + "import random" + ] + }, + { + "cell_type": "code", + "execution_count": 83, + "id": "b7f60f26-0a90-4ea5-b2c2-b5683bda56a6", + "metadata": {}, + "outputs": [], + "source": [ + "env = UnityEnvironment()" + ] + }, + { + "cell_type": "code", + "execution_count": 84, + "id": "5929b410-12c3-4bd9-b984-b2c29a76c3f3", + "metadata": {}, + "outputs": [], + "source": [ + "env.reset()" + ] + }, + { + "cell_type": "code", + "execution_count": 85, + "id": "f108ff09-9f42-4405-add3-6df941c48f8b", + "metadata": { + "scrolled": true, + "tags": [] + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 1. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 1. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n" + ] + }, + { + "ename": "UnityCommunicatorStoppedException", + "evalue": "Communicator has exited.", + "output_type": "error", + "traceback": [ + "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m", + "\u001b[0;31mUnityCommunicatorStoppedException\u001b[0m Traceback (most recent call last)", + "Input \u001b[0;32mIn [85]\u001b[0m, in \u001b[0;36m\u001b[0;34m()\u001b[0m\n\u001b[1;32m 1\u001b[0m \u001b[38;5;28;01mwhile\u001b[39;00m \u001b[38;5;28;01mTrue\u001b[39;00m:\n\u001b[0;32m----> 2\u001b[0m \u001b[43menv\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mstep\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 3\u001b[0m \u001b[38;5;28mprint\u001b[39m(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mStep\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 4\u001b[0m asd \u001b[38;5;241m=\u001b[39m env\u001b[38;5;241m.\u001b[39mget_steps(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mnpc?team=0\u001b[39m\u001b[38;5;124m'\u001b[39m)\n", + "File \u001b[0;32m~/opt/miniforge3/lib/python3.9/site-packages/mlagents_envs/timers.py:305\u001b[0m, in \u001b[0;36mtimed..wrapped\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 303\u001b[0m \u001b[38;5;28;01mdef\u001b[39;00m \u001b[38;5;21mwrapped\u001b[39m(\u001b[38;5;241m*\u001b[39margs, \u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs):\n\u001b[1;32m 304\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m hierarchical_timer(func\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__qualname__\u001b[39m):\n\u001b[0;32m--> 305\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfunc\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", + "File \u001b[0;32m~/opt/miniforge3/lib/python3.9/site-packages/mlagents_envs/environment.py:350\u001b[0m, in \u001b[0;36mUnityEnvironment.step\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 348\u001b[0m outputs \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_communicator\u001b[38;5;241m.\u001b[39mexchange(step_input, \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_poll_process)\n\u001b[1;32m 349\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m outputs \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[0;32m--> 350\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m UnityCommunicatorStoppedException(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mCommunicator has exited.\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 351\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_update_behavior_specs(outputs)\n\u001b[1;32m 352\u001b[0m rl_output \u001b[38;5;241m=\u001b[39m outputs\u001b[38;5;241m.\u001b[39mrl_output\n", + "\u001b[0;31mUnityCommunicatorStoppedException\u001b[0m: Communicator has exited." + ] + } + ], + "source": [ + "while True:\n", + " env.step()\n", + " print(\"Step\")\n", + " asd = env.get_steps('npc?team=0')\n", + " print(asd)\n", + " print(asd[0][0])\n", + " _id = asd[0][0].obs[0][0][0]\n", + " print(_id)\n", + " env.set_action_for_agent('npc?team=0', 0, mlagents_envs.environment.ActionTuple(discrete=np.array([[1, 0, random.randint(0,2)]])))" + ] + }, + { + "cell_type": "code", + "execution_count": 86, + "id": "db100c84-22ab-491b-b68d-4d5c1bbc66a3", + "metadata": {}, + "outputs": [], + "source": [ + "env.close()" + ] + } + ], + "metadata": { + "kernelspec": { + "display_name": "Python 3 (ipykernel)", + "language": "python", + "name": "python3" + }, + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.9.12" + } + }, + "nbformat": 4, + "nbformat_minor": 5 +} -- 2.49.0 From aa7b50554b7b572df14bc2a184b240e4a972b78e Mon Sep 17 00:00:00 2001 From: Andrey Gumirov Date: Sat, 7 May 2022 23:00:24 +0700 Subject: [PATCH 17/17] Added project version --- ProjectSettings/ProjectVersion.txt | 2 ++ 1 file changed, 2 insertions(+) create mode 100644 ProjectSettings/ProjectVersion.txt diff --git a/ProjectSettings/ProjectVersion.txt b/ProjectSettings/ProjectVersion.txt new file mode 100644 index 0000000..bf8f7eb --- /dev/null +++ b/ProjectSettings/ProjectVersion.txt @@ -0,0 +1,2 @@ +m_EditorVersion: 2019.4.36f1 +m_EditorVersionWithRevision: 2019.4.36f1 (660c164b2fc5) -- 2.49.0