From d617ddabe7a25fab5e80e47ba1e4a29f1f341fa8 Mon Sep 17 00:00:00 2001 From: Krazerleo Date: Mon, 9 May 2022 21:31:21 +0700 Subject: [PATCH] Added Animations. Changed inspector view for navpoints. A lot of fixes. Changed project structure. --- Assets/{Defaults.meta => Arts.meta} | 2 +- Assets/{ => Arts}/Materials.meta | 0 Assets/{ => Arts}/Materials/New Material.mat | 0 .../Materials/New Material.mat.meta | 0 Assets/{ => Arts}/Models.meta | 0 Assets/{ => Arts}/Models/Flag model.fbx | 0 Assets/{ => Arts}/Models/Flag model.fbx.meta | 0 Assets/{ => Arts}/Models/Textures.meta | 0 Assets/{ => Arts}/Models/Textures/Flag.jpg | 0 .../{ => Arts}/Models/Textures/Flag.jpg.meta | 0 .../{ => Arts}/SciFiWarriorPBRHPPolyart.meta | 0 .../SciFiWarriorPBRHPPolyart/Animations.meta | 0 .../Animations/AvatarMask.mask | 0 .../Animations/AvatarMask.mask.meta | 0 .../Animations/Die.fbx | 0 .../Animations/Die.fbx.meta | 619 +- .../Animations/Idle_Ducking_ar.fbx | 0 .../Animations/Idle_Ducking_ar.fbx.meta | 623 +- .../Animations/Idle_Shoot_ar.fbx | 0 .../Animations/Idle_Shoot_ar.fbx.meta | 0 .../Animations/Idle_guard_ar.fbx | 0 .../Animations/Idle_guard_ar.fbx.meta | 0 .../Animations/Idle_gunMiddle_ar.fbx | 0 .../Animations/Idle_gunMiddle_ar.fbx.meta | 619 +- .../Animations/Jump.fbx | 0 .../Animations/Jump.fbx.meta | 0 .../Animations/Reload.fbx | 0 .../Animations/Reload.fbx.meta | 0 .../Animations/Run_guard_AR.fbx | 0 .../Animations/Run_guard_AR.fbx.meta | 0 .../Animations/Run_gunMiddle_AR.fbx | 0 .../Animations/Run_gunMiddle_AR.fbx.meta | 0 .../Animations/Shoot_AutoShot_AR.fbx | 0 .../Animations/Shoot_AutoShot_AR.fbx.meta | 0 .../Animations/Shoot_BurstShot_AR.fbx | 0 .../Animations/Shoot_BurstShot_AR.fbx.meta | 0 .../Animations/Shoot_SingleShot_AR.fbx | 0 .../Animations/Shoot_SingleShot_AR.fbx.meta | 0 .../Animations/WalkBack_Shoot_ar.fbx | 0 .../Animations/WalkBack_Shoot_ar.fbx.meta | 0 .../Animations/WalkFront_Shoot_ar.fbx | 0 .../Animations/WalkFront_Shoot_ar.fbx.meta | 0 .../Animations/WalkLeft_Shoot_ar.fbx | 0 .../Animations/WalkLeft_Shoot_ar.fbx.meta | 0 .../Animations/WalkRight_Shoot_ar.fbx | 0 .../Animations/WalkRight_Shoot_ar.fbx.meta | 0 .../SciFiWarriorPBRHPPolyart/Animators.meta | 0 .../Animators/SciFiWarrior.controller | 93 + .../Animators/SciFiWarrior.controller.meta | 0 .../Animators/SciFiWarriorOur.controller | 0 .../Animators/SciFiWarriorOur.controller.meta | 0 .../SciFiWarriorPBRHPPolyart/Materials.meta | 0 .../SciFiWarriorPBRHPPolyart/Materials/HP.mat | 0 .../Materials/HP.mat.meta | 0 .../Materials/PBR.mat | 0 .../Materials/PBR.mat.meta | 0 .../Materials/Polyart.mat | 0 .../Materials/Polyart.mat.meta | 0 .../Materials/Skybox_Mat.mat | 178 +- .../Materials/Skybox_Mat.mat.meta | 0 .../Materials/Stage.mat | 0 .../Materials/Stage.mat.meta | 0 .../SciFiWarriorPBRHPPolyart/Mesh.meta | 0 .../Mesh/PBR_HP_Mesh.fbx | 0 .../Mesh/PBR_HP_Mesh.fbx.meta | 0 .../Mesh/Polyart_Mesh.fbx | 0 .../Mesh/Polyart_Mesh.fbx.meta | 0 .../SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx | 0 .../Mesh/Stage.fbx.meta | 0 .../SciFiWarriorPBRHPPolyart/Prefabs.meta | 0 .../Prefabs/HPCharacter.prefab | 0 .../Prefabs/HPCharacter.prefab.meta | 0 .../Prefabs/PBRCharacter.prefab | 0 .../Prefabs/PBRCharacter.prefab.meta | 0 .../Prefabs/PolyartCharacter.prefab | 102 + .../Prefabs/PolyartCharacter.prefab.meta | 0 .../SciFiWarriorPBRHPPolyart/Scene.meta | 0 .../Scene/DemoScene01.unity | 0 .../Scene/DemoScene01.unity.meta | 0 .../SciFiWarriorPBRHPPolyart/Textures.meta | 0 .../Textures/HP_Albedo.png | 0 .../Textures/HP_Albedo.png.meta | 0 .../Textures/PBR_Albedo.png | 0 .../Textures/PBR_Albedo.png.meta | 0 .../Textures/PBR_Free_AO.png | 0 .../Textures/PBR_Free_AO.png.meta | 0 .../Textures/PBR_Free_EM.png | 0 .../Textures/PBR_Free_EM.png.meta | 0 .../Textures/PBR_Free_MS.png | 0 .../Textures/PBR_Free_MS.png.meta | 0 .../Textures/PBR_Free_NM.png | 0 .../Textures/PBR_Free_NM.png.meta | 0 .../Textures/Polyart.png | 0 .../Textures/Polyart.png.meta | 0 .../Textures/PolyartEmission.png | 0 .../Textures/PolyartEmission.png.meta | 0 .../SciFiWarriorPBRHPPolyart/promotion.jpg | 0 .../promotion.jpg.meta | 0 Assets/Defaults/mat_proBuilder.mat | 77 - Assets/Inputs/InputSystem.inputsettings.asset | 35 + .../InputSystem.inputsettings.asset.meta | 8 + Assets/Inputs/ThirdPersonViewInput.cs | 354 + Assets/Inputs/ThirdPersonViewInput.cs.meta | 11 + .../Inputs/ThirdPersonViewInput.inputactions | 166 + .../ThirdPersonViewInput.inputactions.meta | 14 + .../ML-Agents/Timers/DemoScene01_timers.json | 1 - .../Timers/dont touch me plz_timers.json | 1 - Assets/Prefabs/CharacterPrefabs.meta | 8 + .../Prefabs/{ => CharacterPrefabs}/Bot.prefab | 0 .../{ => CharacterPrefabs}/Bot.prefab.meta | 0 .../Prefabs/CharacterPrefabs/Character.prefab | 2639 ++ .../CharacterPrefabs/Character.prefab.meta} | 4 +- .../{ => CharacterPrefabs}/Player.prefab | 0 .../{ => CharacterPrefabs}/Player.prefab.meta | 0 Assets/Prefabs/EntityPrefabs.meta | 8 + .../{ => EntityPrefabs}/Flag zone.prefab | 0 .../{ => EntityPrefabs}/Flag zone.prefab.meta | 0 .../EntityPrefabs/navPointPrefab.prefab | 49 + .../EntityPrefabs/navPointPrefab.prefab.meta | 7 + Assets/Prefabs/PickupPrefabs.meta | 8 + .../Prefabs/{ => PickupPrefabs}/ammo.prefab | 0 .../{ => PickupPrefabs}/ammo.prefab.meta | 0 .../Prefabs/{ => PickupPrefabs}/armour.prefab | 0 .../{ => PickupPrefabs}/armour.prefab.meta | 0 .../Prefabs/{ => PickupPrefabs}/health.prefab | 0 .../{ => PickupPrefabs}/health.prefab.meta | 0 Assets/Resources.meta | 8 + Assets/Resources/BillingMode.json | 1 + .../BillingMode.json.meta} | 2 +- Assets/Scenes/AnimationTestScene.unity | 1060 + ...ity.meta => AnimationTestScene.unity.meta} | 2 +- .../Greatest_map_ever/Greatest_map_ever.unity | 166 +- .../Scenes/{First try.unity => Map#1.unity} | 0 ...{First try.unity.meta => Map#1.unity.meta} | 0 Assets/Scenes/dont touch me plz.unity | 460 - Assets/Scripts/Animators.meta | 3 + Assets/Scripts/Animators/Kirill Animator.meta | 3 + .../CustomCharacterController.cs} | 508 +- .../CustomCharacterController.cs.meta} | 0 .../Animators/Kirill Animator/Models.cs | 67 + .../Kirill Animator/Models.cs.meta} | 0 Assets/Scripts/Animators/Leonid Animator.meta | 3 + .../Leonid Animator/AnimatorHandler.cs | 94 + .../Leonid Animator/AnimatorHandler.cs.meta | 3 + .../CharacterAnimator.controller} | 1289 +- .../CharacterAnimator.controller.meta} | 4 +- .../Leonid Animator/CharacterLocomotion.cs | 105 + .../CharacterLocomotion.cs.meta | 3 + .../Animators/Leonid Animator/InputHandler.cs | 77 + .../Leonid Animator/InputHandler.cs.meta | 3 + .../Animators/Leonid Animator/LowerBody.mask | 139 + .../Leonid Animator/LowerBody.mask.meta | 8 + .../Animators/Leonid Animator/UpperBody.mask | 139 + .../Leonid Animator/UpperBody.mask.meta | 8 + Assets/Scripts/CameraScripts.meta | 3 + Assets/Scripts/CameraScripts/CameraHandler.cs | 61 + .../CameraScripts/CameraHandler.cs.meta | 3 + Assets/Scripts/Character/NPC.cs | 1 + Assets/Scripts/Character/scr_Models.cs | 75 - Assets/Scripts/Misc/NavPoint.cs | 9 +- ...eaponController.cs => WeaponController.cs} | 93 +- ...oller.cs.meta => WeaponController.cs.meta} | 0 ProjectSettings/EditorBuildSettings.asset | 4 +- ProjectSettings/ProjectVersion.txt | 2 + ProjectSettings/TagManager.asset | 2 +- test-ml-agents.ipynb | 20623 ++++++++++++++++ 166 files changed, 28460 insertions(+), 2197 deletions(-) rename Assets/{Defaults.meta => Arts.meta} (77%) mode change 100755 => 100644 rename Assets/{ => Arts}/Materials.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Materials/New Material.mat (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Materials/New Material.mat.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Models.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Models/Flag model.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Models/Flag model.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Models/Textures.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Models/Textures/Flag.jpg (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/Models/Textures/Flag.jpg.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Die.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Die.fbx.meta (79%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx.meta (79%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx.meta (79%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animators.meta (100%) mode change 100755 => 100644 create mode 100644 Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/HP.mat (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/HP.mat.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/PBR.mat (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/PBR.mat.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat (96%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/Stage.mat (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Materials/Stage.mat.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Mesh.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Prefabs.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab (96%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Scene.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/Polyart.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/Polyart.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png.meta (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/promotion.jpg (100%) mode change 100755 => 100644 rename Assets/{ => Arts}/SciFiWarriorPBRHPPolyart/promotion.jpg.meta (100%) mode change 100755 => 100644 delete mode 100755 Assets/Defaults/mat_proBuilder.mat create mode 100644 Assets/Inputs/InputSystem.inputsettings.asset create mode 100644 Assets/Inputs/InputSystem.inputsettings.asset.meta create mode 100644 Assets/Inputs/ThirdPersonViewInput.cs create mode 100644 Assets/Inputs/ThirdPersonViewInput.cs.meta create mode 100644 Assets/Inputs/ThirdPersonViewInput.inputactions create mode 100644 Assets/Inputs/ThirdPersonViewInput.inputactions.meta delete mode 100755 Assets/ML-Agents/Timers/DemoScene01_timers.json delete mode 100755 Assets/ML-Agents/Timers/dont touch me plz_timers.json create mode 100644 Assets/Prefabs/CharacterPrefabs.meta rename Assets/Prefabs/{ => CharacterPrefabs}/Bot.prefab (100%) rename Assets/Prefabs/{ => CharacterPrefabs}/Bot.prefab.meta (100%) mode change 100755 => 100644 create mode 100644 Assets/Prefabs/CharacterPrefabs/Character.prefab rename Assets/{ML-Agents/Timers/dont touch me plz_timers.json.meta => Prefabs/CharacterPrefabs/Character.prefab.meta} (62%) mode change 100755 => 100644 rename Assets/Prefabs/{ => CharacterPrefabs}/Player.prefab (100%) mode change 100755 => 100644 rename Assets/Prefabs/{ => CharacterPrefabs}/Player.prefab.meta (100%) create mode 100644 Assets/Prefabs/EntityPrefabs.meta rename Assets/Prefabs/{ => EntityPrefabs}/Flag zone.prefab (100%) mode change 100755 => 100644 rename Assets/Prefabs/{ => EntityPrefabs}/Flag zone.prefab.meta (100%) mode change 100755 => 100644 create mode 100644 Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab create mode 100644 Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab.meta create mode 100644 Assets/Prefabs/PickupPrefabs.meta rename Assets/Prefabs/{ => PickupPrefabs}/ammo.prefab (100%) mode change 100755 => 100644 rename Assets/Prefabs/{ => PickupPrefabs}/ammo.prefab.meta (100%) mode change 100755 => 100644 rename Assets/Prefabs/{ => PickupPrefabs}/armour.prefab (100%) mode change 100755 => 100644 rename Assets/Prefabs/{ => PickupPrefabs}/armour.prefab.meta (100%) mode change 100755 => 100644 rename Assets/Prefabs/{ => PickupPrefabs}/health.prefab (100%) mode change 100755 => 100644 rename Assets/Prefabs/{ => PickupPrefabs}/health.prefab.meta (100%) mode change 100755 => 100644 create mode 100644 Assets/Resources.meta create mode 100644 Assets/Resources/BillingMode.json rename Assets/{ML-Agents/Timers/DemoScene01_timers.json.meta => Resources/BillingMode.json.meta} (75%) mode change 100755 => 100644 create mode 100644 Assets/Scenes/AnimationTestScene.unity rename Assets/Scenes/{dont touch me plz.unity.meta => AnimationTestScene.unity.meta} (74%) mode change 100755 => 100644 rename Assets/Scenes/{First try.unity => Map#1.unity} (100%) mode change 100755 => 100644 rename Assets/Scenes/{First try.unity.meta => Map#1.unity.meta} (100%) mode change 100755 => 100644 delete mode 100755 Assets/Scenes/dont touch me plz.unity create mode 100644 Assets/Scripts/Animators.meta create mode 100644 Assets/Scripts/Animators/Kirill Animator.meta rename Assets/Scripts/{Character/scr_CharacterController.cs => Animators/Kirill Animator/CustomCharacterController.cs} (95%) mode change 100755 => 100644 rename Assets/Scripts/{Character/scr_CharacterController.cs.meta => Animators/Kirill Animator/CustomCharacterController.cs.meta} (100%) mode change 100755 => 100644 create mode 100644 Assets/Scripts/Animators/Kirill Animator/Models.cs rename Assets/Scripts/{Character/scr_Models.cs.meta => Animators/Kirill Animator/Models.cs.meta} (100%) mode change 100755 => 100644 create mode 100644 Assets/Scripts/Animators/Leonid Animator.meta create mode 100644 Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs create mode 100644 Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs.meta rename Assets/{SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller => Scripts/Animators/Leonid Animator/CharacterAnimator.controller} (56%) mode change 100755 => 100644 rename Assets/{Defaults/mat_proBuilder.mat.meta => Scripts/Animators/Leonid Animator/CharacterAnimator.controller.meta} (64%) mode change 100755 => 100644 create mode 100644 Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs create mode 100644 Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs.meta create mode 100644 Assets/Scripts/Animators/Leonid Animator/InputHandler.cs create mode 100644 Assets/Scripts/Animators/Leonid Animator/InputHandler.cs.meta create mode 100644 Assets/Scripts/Animators/Leonid Animator/LowerBody.mask create mode 100644 Assets/Scripts/Animators/Leonid Animator/LowerBody.mask.meta create mode 100644 Assets/Scripts/Animators/Leonid Animator/UpperBody.mask create mode 100644 Assets/Scripts/Animators/Leonid Animator/UpperBody.mask.meta create mode 100644 Assets/Scripts/CameraScripts.meta create mode 100644 Assets/Scripts/CameraScripts/CameraHandler.cs create mode 100644 Assets/Scripts/CameraScripts/CameraHandler.cs.meta delete mode 100755 Assets/Scripts/Character/scr_Models.cs rename Assets/Scripts/Weapons/{scr_WeaponController.cs => WeaponController.cs} (96%) mode change 100755 => 100644 rename Assets/Scripts/Weapons/{scr_WeaponController.cs.meta => WeaponController.cs.meta} (100%) mode change 100755 => 100644 create mode 100644 ProjectSettings/ProjectVersion.txt create mode 100644 test-ml-agents.ipynb diff --git a/Assets/Defaults.meta b/Assets/Arts.meta old mode 100755 new mode 100644 similarity index 77% rename from Assets/Defaults.meta rename to Assets/Arts.meta index f41c53d..a5b76ea --- a/Assets/Defaults.meta +++ b/Assets/Arts.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: b183268306ddaac4b875d5d435faa50b +guid: 0d3fef053a0198e44a77e5bbb1c9e6ba folderAsset: yes DefaultImporter: externalObjects: {} diff --git a/Assets/Materials.meta b/Assets/Arts/Materials.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Materials.meta rename to Assets/Arts/Materials.meta diff --git a/Assets/Materials/New Material.mat b/Assets/Arts/Materials/New Material.mat old mode 100755 new mode 100644 similarity index 100% rename from Assets/Materials/New Material.mat rename to Assets/Arts/Materials/New Material.mat diff --git a/Assets/Materials/New Material.mat.meta b/Assets/Arts/Materials/New Material.mat.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Materials/New Material.mat.meta rename to Assets/Arts/Materials/New Material.mat.meta diff --git a/Assets/Models.meta b/Assets/Arts/Models.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Models.meta rename to Assets/Arts/Models.meta diff --git a/Assets/Models/Flag model.fbx b/Assets/Arts/Models/Flag model.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/Models/Flag model.fbx rename to Assets/Arts/Models/Flag model.fbx diff --git a/Assets/Models/Flag model.fbx.meta b/Assets/Arts/Models/Flag model.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Models/Flag model.fbx.meta rename to Assets/Arts/Models/Flag model.fbx.meta diff --git a/Assets/Models/Textures.meta b/Assets/Arts/Models/Textures.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Models/Textures.meta rename to Assets/Arts/Models/Textures.meta diff --git a/Assets/Models/Textures/Flag.jpg b/Assets/Arts/Models/Textures/Flag.jpg old mode 100755 new mode 100644 similarity index 100% rename from Assets/Models/Textures/Flag.jpg rename to Assets/Arts/Models/Textures/Flag.jpg diff --git a/Assets/Models/Textures/Flag.jpg.meta b/Assets/Arts/Models/Textures/Flag.jpg.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Models/Textures/Flag.jpg.meta rename to Assets/Arts/Models/Textures/Flag.jpg.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/AvatarMask.mask.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Die.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Die.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Die.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Die.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Die.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Die.fbx.meta old mode 100755 new mode 100644 similarity index 79% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Die.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Die.fbx.meta index 7d5f536..5e3e975 --- a/Assets/SciFiWarriorPBRHPPolyart/Animations/Die.fbx.meta +++ b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Die.fbx.meta @@ -1,160 +1,458 @@ fileFormatVersion: 2 guid: d406f8f3cbe268f4e9d0234d45cca60c ModelImporter: - serializedVersion: 23 - fileIDToRecycleName: - 100000: Arm1 - 100002: ArmPlacement_Left - 100004: ArmPlacement_Right - 100006: ArmPlacement_Upper - 100008: ArmPosition_Left - 100010: ArmPosition_Right - 100012: AssaultRifle - 100014: BackPack - 100016: Backpack1 - 100018: Body1 - 100020: Chest - 100022: //RootNode - 100024: Foot_Left - 100026: Foot_Right - 100028: Hand_Left - 100030: Hand_Right - 100032: Head - 100034: head1 - 100036: Headgear_Left - 100038: Headgear_Right - 100040: Hips - 100042: Index_Distal_Left - 100044: Index_Distal_Right - 100046: Index_Intermediate_Left - 100048: Index_Intermediate_Right - 100050: Index_Proximal_Left - 100052: Index_Proximal_Right - 100054: Leg1 - 100056: LowerArm_Left - 100058: LowerArm_Right - 100060: LowerLeg_Left - 100062: LowerLeg_Right - 100064: magazine_Right - 100066: Neck - 100068: RestOfFingers_Distal_Left - 100070: RestOfFingers_Distal_Right - 100072: RestOfFingers_Intermediate_Left - 100074: RestOfFingers_Intermediate_Right - 100076: RestOfFingers_Proximal_Left - 100078: RestOfFingers_Proximal_Right - 100080: Shoulder_Left - 100082: Shoulder_Right - 100084: ShoulderPadBlade_Left - 100086: ShoulderPadBlade_Right - 100088: ShoulderPadBody_Left - 100090: ShoulderPadBody_Right - 100092: ShoulderPadCTRL_Left - 100094: ShoulderPadCTRL_Right - 100096: Spine - 100098: Thumb_Distal_Left - 100100: Thumb_Distal_Right - 100102: Thumb_Intermediate_Left - 100104: Thumb_Intermediate_Right - 100106: Thumb_Proximal_Left - 100108: Thumb_Proximal_Right - 100110: Toe_Left - 100112: Toe_Right - 100114: Toetip_Left - 100116: Toetip_Right - 100118: Trigger_Right - 100120: UpperArm_Left - 100122: UpperArm_Right - 100124: UpperLeg_Left - 100126: UpperLeg_Right - 400000: Arm1 - 400002: ArmPlacement_Left - 400004: ArmPlacement_Right - 400006: ArmPlacement_Upper - 400008: ArmPosition_Left - 400010: ArmPosition_Right - 400012: AssaultRifle - 400014: BackPack - 400016: Backpack1 - 400018: Body1 - 400020: Chest - 400022: //RootNode - 400024: Foot_Left - 400026: Foot_Right - 400028: Hand_Left - 400030: Hand_Right - 400032: Head - 400034: head1 - 400036: Headgear_Left - 400038: Headgear_Right - 400040: Hips - 400042: Index_Distal_Left - 400044: Index_Distal_Right - 400046: Index_Intermediate_Left - 400048: Index_Intermediate_Right - 400050: Index_Proximal_Left - 400052: Index_Proximal_Right - 400054: Leg1 - 400056: LowerArm_Left - 400058: LowerArm_Right - 400060: LowerLeg_Left - 400062: LowerLeg_Right - 400064: magazine_Right - 400066: Neck - 400068: RestOfFingers_Distal_Left - 400070: RestOfFingers_Distal_Right - 400072: RestOfFingers_Intermediate_Left - 400074: RestOfFingers_Intermediate_Right - 400076: RestOfFingers_Proximal_Left - 400078: RestOfFingers_Proximal_Right - 400080: Shoulder_Left - 400082: Shoulder_Right - 400084: ShoulderPadBlade_Left - 400086: ShoulderPadBlade_Right - 400088: ShoulderPadBody_Left - 400090: ShoulderPadBody_Right - 400092: ShoulderPadCTRL_Left - 400094: ShoulderPadCTRL_Right - 400096: Spine - 400098: Thumb_Distal_Left - 400100: Thumb_Distal_Right - 400102: Thumb_Intermediate_Left - 400104: Thumb_Intermediate_Right - 400106: Thumb_Proximal_Left - 400108: Thumb_Proximal_Right - 400110: Toe_Left - 400112: Toe_Right - 400114: Toetip_Left - 400116: Toetip_Right - 400118: Trigger_Right - 400120: UpperArm_Left - 400122: UpperArm_Right - 400124: UpperLeg_Left - 400126: UpperLeg_Right - 2100000: w_h_01 - 2100002: w_b_01 - 2100004: w_bp_01 - 2100006: w_a_01 - 2100008: w_f_01 - 2100010: w_w_01 - 4300000: head1 - 4300002: Body1 - 4300004: Backpack1 - 4300006: Arm1 - 4300008: Leg1 - 4300010: AssaultRifle - 7400000: Die - 9500000: //RootNode - 13700000: Arm1 - 13700002: AssaultRifle - 13700004: Backpack1 - 13700006: Body1 - 13700008: head1 - 13700010: Leg1 - 2186277476908879412: ImportLogs + serializedVersion: 19301 + internalIDToNameTable: + - first: + 1: 100000 + second: Arm1 + - first: + 1: 100002 + second: ArmPlacement_Left + - first: + 1: 100004 + second: ArmPlacement_Right + - first: + 1: 100006 + second: ArmPlacement_Upper + - first: + 1: 100008 + second: ArmPosition_Left + - first: + 1: 100010 + second: ArmPosition_Right + - first: + 1: 100012 + second: AssaultRifle + - first: + 1: 100014 + second: BackPack + - first: + 1: 100016 + second: Backpack1 + - first: + 1: 100018 + second: Body1 + - first: + 1: 100020 + second: Chest + - first: + 1: 100022 + second: //RootNode + - first: + 1: 100024 + second: Foot_Left + - first: + 1: 100026 + second: Foot_Right + - first: + 1: 100028 + second: Hand_Left + - first: + 1: 100030 + second: Hand_Right + - first: + 1: 100032 + second: Head + - first: + 1: 100034 + second: head1 + - first: + 1: 100036 + second: Headgear_Left + - first: + 1: 100038 + second: Headgear_Right + - first: + 1: 100040 + second: Hips + - first: + 1: 100042 + second: Index_Distal_Left + - first: + 1: 100044 + second: Index_Distal_Right + - first: + 1: 100046 + second: Index_Intermediate_Left + - first: + 1: 100048 + second: Index_Intermediate_Right + - first: + 1: 100050 + second: Index_Proximal_Left + - first: + 1: 100052 + second: Index_Proximal_Right + - first: + 1: 100054 + second: Leg1 + - first: + 1: 100056 + second: LowerArm_Left + - first: + 1: 100058 + second: LowerArm_Right + - first: + 1: 100060 + second: LowerLeg_Left + - first: + 1: 100062 + second: LowerLeg_Right + - first: + 1: 100064 + second: magazine_Right + - first: + 1: 100066 + second: Neck + - first: + 1: 100068 + second: RestOfFingers_Distal_Left + - first: + 1: 100070 + second: RestOfFingers_Distal_Right + - first: + 1: 100072 + second: RestOfFingers_Intermediate_Left + - first: + 1: 100074 + second: RestOfFingers_Intermediate_Right + - first: + 1: 100076 + second: RestOfFingers_Proximal_Left + - first: + 1: 100078 + second: RestOfFingers_Proximal_Right + - first: + 1: 100080 + second: Shoulder_Left + - first: + 1: 100082 + second: Shoulder_Right + - first: + 1: 100084 + second: ShoulderPadBlade_Left + - first: + 1: 100086 + second: ShoulderPadBlade_Right + - first: + 1: 100088 + second: ShoulderPadBody_Left + - first: + 1: 100090 + second: ShoulderPadBody_Right + - first: + 1: 100092 + second: ShoulderPadCTRL_Left + - first: + 1: 100094 + second: ShoulderPadCTRL_Right + - first: + 1: 100096 + second: Spine + - first: + 1: 100098 + second: Thumb_Distal_Left + - first: + 1: 100100 + second: Thumb_Distal_Right + - first: + 1: 100102 + second: Thumb_Intermediate_Left + - first: + 1: 100104 + second: Thumb_Intermediate_Right + - first: + 1: 100106 + second: Thumb_Proximal_Left + - first: + 1: 100108 + second: Thumb_Proximal_Right + - first: + 1: 100110 + second: Toe_Left + - first: + 1: 100112 + second: Toe_Right + - first: + 1: 100114 + second: Toetip_Left + - first: + 1: 100116 + second: Toetip_Right + - first: + 1: 100118 + second: Trigger_Right + - first: + 1: 100120 + second: UpperArm_Left + - first: + 1: 100122 + second: UpperArm_Right + - first: + 1: 100124 + second: UpperLeg_Left + - first: + 1: 100126 + second: UpperLeg_Right + - first: + 4: 400000 + second: Arm1 + - first: + 4: 400002 + second: ArmPlacement_Left + - first: + 4: 400004 + second: ArmPlacement_Right + - first: + 4: 400006 + second: ArmPlacement_Upper + - first: + 4: 400008 + second: ArmPosition_Left + - first: + 4: 400010 + second: ArmPosition_Right + - first: + 4: 400012 + second: AssaultRifle + - first: + 4: 400014 + second: BackPack + - first: + 4: 400016 + second: Backpack1 + - first: + 4: 400018 + second: Body1 + - first: + 4: 400020 + second: Chest + - first: + 4: 400022 + second: //RootNode + - first: + 4: 400024 + second: Foot_Left + - first: + 4: 400026 + second: Foot_Right + - first: + 4: 400028 + second: Hand_Left + - first: + 4: 400030 + second: Hand_Right + - first: + 4: 400032 + second: Head + - first: + 4: 400034 + second: head1 + - first: + 4: 400036 + second: Headgear_Left + - first: + 4: 400038 + second: Headgear_Right + - first: + 4: 400040 + second: Hips + - first: + 4: 400042 + second: Index_Distal_Left + - first: + 4: 400044 + second: Index_Distal_Right + - first: + 4: 400046 + second: Index_Intermediate_Left + - first: + 4: 400048 + second: Index_Intermediate_Right + - first: + 4: 400050 + second: Index_Proximal_Left + - first: + 4: 400052 + second: Index_Proximal_Right + - first: + 4: 400054 + second: Leg1 + - first: + 4: 400056 + second: LowerArm_Left + - first: + 4: 400058 + second: LowerArm_Right + - first: + 4: 400060 + second: LowerLeg_Left + - first: + 4: 400062 + second: LowerLeg_Right + - first: + 4: 400064 + second: magazine_Right + - first: + 4: 400066 + second: Neck + - first: + 4: 400068 + second: RestOfFingers_Distal_Left + - first: + 4: 400070 + second: RestOfFingers_Distal_Right + - first: + 4: 400072 + second: RestOfFingers_Intermediate_Left + - first: + 4: 400074 + second: RestOfFingers_Intermediate_Right + - first: + 4: 400076 + second: RestOfFingers_Proximal_Left + - first: + 4: 400078 + second: RestOfFingers_Proximal_Right + - first: + 4: 400080 + second: Shoulder_Left + - first: + 4: 400082 + second: Shoulder_Right + - first: + 4: 400084 + second: ShoulderPadBlade_Left + - first: + 4: 400086 + second: ShoulderPadBlade_Right + - first: + 4: 400088 + second: ShoulderPadBody_Left + - first: + 4: 400090 + second: ShoulderPadBody_Right + - first: + 4: 400092 + second: ShoulderPadCTRL_Left + - first: + 4: 400094 + second: ShoulderPadCTRL_Right + - first: + 4: 400096 + second: Spine + - first: + 4: 400098 + second: Thumb_Distal_Left + - first: + 4: 400100 + second: Thumb_Distal_Right + - first: + 4: 400102 + second: Thumb_Intermediate_Left + - first: + 4: 400104 + second: Thumb_Intermediate_Right + - first: + 4: 400106 + second: Thumb_Proximal_Left + - first: + 4: 400108 + second: Thumb_Proximal_Right + - first: + 4: 400110 + second: Toe_Left + - first: + 4: 400112 + second: Toe_Right + - first: + 4: 400114 + second: Toetip_Left + - first: + 4: 400116 + second: Toetip_Right + - first: + 4: 400118 + second: Trigger_Right + - first: + 4: 400120 + second: UpperArm_Left + - first: + 4: 400122 + second: UpperArm_Right + - first: + 4: 400124 + second: UpperLeg_Left + - first: + 4: 400126 + second: UpperLeg_Right + - first: + 21: 2100000 + second: w_h_01 + - first: + 21: 2100002 + second: w_b_01 + - first: + 21: 2100004 + second: w_bp_01 + - first: + 21: 2100006 + second: w_a_01 + - first: + 21: 2100008 + second: w_f_01 + - first: + 21: 2100010 + second: w_w_01 + - first: + 43: 4300000 + second: head1 + - first: + 43: 4300002 + second: Body1 + - first: + 43: 4300004 + second: Backpack1 + - first: + 43: 4300006 + second: Arm1 + - first: + 43: 4300008 + second: Leg1 + - first: + 43: 4300010 + second: AssaultRifle + - first: + 74: 7400000 + second: Die + - first: + 95: 9500000 + second: //RootNode + - first: + 137: 13700000 + second: Arm1 + - first: + 137: 13700002 + second: AssaultRifle + - first: + 137: 13700004 + second: Backpack1 + - first: + 137: 13700006 + second: Body1 + - first: + 137: 13700008 + second: head1 + - first: + 137: 13700010 + second: Leg1 + - first: + 41386430: 2186277476908879412 + second: ImportLogs externalObjects: {} materials: - importMaterials: 1 + materialImportMode: 1 materialName: 0 materialSearch: 1 materialLocation: 1 @@ -183,6 +481,7 @@ ModelImporter: - serializedVersion: 16 name: Die takeName: Take 001 + internalID: 0 firstFrame: 0 lastFrame: 32 wrapMode: 0 @@ -343,25 +642,28 @@ ModelImporter: meshCompression: 0 addColliders: 0 useSRGBMaterialColor: 1 + sortHierarchyByName: 1 importVisibility: 1 importBlendShapes: 1 importCameras: 1 importLights: 1 + fileIdsGeneration: 1 swapUVChannels: 0 generateSecondaryUV: 0 useFileUnits: 1 - optimizeMeshForGPU: 1 keepQuads: 0 weldVertices: 1 preserveHierarchy: 0 + skinWeightsMode: 0 + maxBonesPerVertex: 4 + minBoneWeight: 0.001 + meshOptimizationFlags: -1 indexFormat: 0 secondaryUVAngleDistortion: 8 secondaryUVAreaDistortion: 15.000001 secondaryUVHardAngle: 88 secondaryUVPackMargin: 4 useFileScale: 1 - previousCalculatedGlobalScale: 1 - hasPreviousCalculatedGlobalScale: 1 tangentSpace: normalSmoothAngle: 60 normalImportMode: 0 @@ -370,10 +672,10 @@ ModelImporter: legacyComputeAllNormalsFromSmoothingGroupsWhenMeshHasBlendShapes: 0 blendShapeNormalImportMode: 1 normalSmoothingSource: 0 + referencedClips: [] importAnimation: 1 - copyAvatar: 0 humanDescription: - serializedVersion: 2 + serializedVersion: 3 human: - boneName: Hips humanName: Hips @@ -1015,13 +1317,16 @@ ModelImporter: armStretch: 0.05 legStretch: 0.05 feetSpacing: 0 + globalScale: 1 rootMotionBoneName: hasTranslationDoF: 0 hasExtraRoot: 1 skeletonHasParents: 1 lastHumanDescriptionAvatarSource: {instanceID: 0} + autoGenerateAvatarMappingIfUnspecified: 1 animationType: 3 humanoidOversampling: 1 + avatarSetup: 1 additionalBone: 0 userData: assetBundleName: diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx.meta old mode 100755 new mode 100644 similarity index 79% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx.meta index f587e0e..b9863d4 --- a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx.meta +++ b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Ducking_ar.fbx.meta @@ -1,160 +1,458 @@ fileFormatVersion: 2 guid: 3ad7c5979f6586d4a9532a55492a0ebe ModelImporter: - serializedVersion: 23 - fileIDToRecycleName: - 100000: Arm1 - 100002: ArmPlacement_Left - 100004: ArmPlacement_Right - 100006: ArmPlacement_Upper - 100008: ArmPosition_Left - 100010: ArmPosition_Right - 100012: AssaultRifle - 100014: BackPack - 100016: Backpack1 - 100018: Body1 - 100020: Chest - 100022: Foot_Left - 100024: Foot_Right - 100026: Hand_Left - 100028: Hand_Right - 100030: Head - 100032: head1 - 100034: Headgear_Left - 100036: Headgear_Right - 100038: Hips - 100040: //RootNode - 100042: Index_Distal_Left - 100044: Index_Distal_Right - 100046: Index_Intermediate_Left - 100048: Index_Intermediate_Right - 100050: Index_Proximal_Left - 100052: Index_Proximal_Right - 100054: Leg1 - 100056: LowerArm_Left - 100058: LowerArm_Right - 100060: LowerLeg_Left - 100062: LowerLeg_Right - 100064: magazine_Right - 100066: Neck - 100068: RestOfFingers_Distal_Left - 100070: RestOfFingers_Distal_Right - 100072: RestOfFingers_Intermediate_Left - 100074: RestOfFingers_Intermediate_Right - 100076: RestOfFingers_Proximal_Left - 100078: RestOfFingers_Proximal_Right - 100080: Shoulder_Left - 100082: Shoulder_Right - 100084: ShoulderPadBlade_Left - 100086: ShoulderPadBlade_Right - 100088: ShoulderPadBody_Left - 100090: ShoulderPadBody_Right - 100092: ShoulderPadCTRL_Left - 100094: ShoulderPadCTRL_Right - 100096: Spine - 100098: Thumb_Distal_Left - 100100: Thumb_Distal_Right - 100102: Thumb_Intermediate_Left - 100104: Thumb_Intermediate_Right - 100106: Thumb_Proximal_Left - 100108: Thumb_Proximal_Right - 100110: Toe_Left - 100112: Toe_Right - 100114: Toetip_Left - 100116: Toetip_Right - 100118: Trigger_Right - 100120: UpperArm_Left - 100122: UpperArm_Right - 100124: UpperLeg_Left - 100126: UpperLeg_Right - 400000: Arm1 - 400002: ArmPlacement_Left - 400004: ArmPlacement_Right - 400006: ArmPlacement_Upper - 400008: ArmPosition_Left - 400010: ArmPosition_Right - 400012: AssaultRifle - 400014: BackPack - 400016: Backpack1 - 400018: Body1 - 400020: Chest - 400022: Foot_Left - 400024: Foot_Right - 400026: Hand_Left - 400028: Hand_Right - 400030: Head - 400032: head1 - 400034: Headgear_Left - 400036: Headgear_Right - 400038: Hips - 400040: //RootNode - 400042: Index_Distal_Left - 400044: Index_Distal_Right - 400046: Index_Intermediate_Left - 400048: Index_Intermediate_Right - 400050: Index_Proximal_Left - 400052: Index_Proximal_Right - 400054: Leg1 - 400056: LowerArm_Left - 400058: LowerArm_Right - 400060: LowerLeg_Left - 400062: LowerLeg_Right - 400064: magazine_Right - 400066: Neck - 400068: RestOfFingers_Distal_Left - 400070: RestOfFingers_Distal_Right - 400072: RestOfFingers_Intermediate_Left - 400074: RestOfFingers_Intermediate_Right - 400076: RestOfFingers_Proximal_Left - 400078: RestOfFingers_Proximal_Right - 400080: Shoulder_Left - 400082: Shoulder_Right - 400084: ShoulderPadBlade_Left - 400086: ShoulderPadBlade_Right - 400088: ShoulderPadBody_Left - 400090: ShoulderPadBody_Right - 400092: ShoulderPadCTRL_Left - 400094: ShoulderPadCTRL_Right - 400096: Spine - 400098: Thumb_Distal_Left - 400100: Thumb_Distal_Right - 400102: Thumb_Intermediate_Left - 400104: Thumb_Intermediate_Right - 400106: Thumb_Proximal_Left - 400108: Thumb_Proximal_Right - 400110: Toe_Left - 400112: Toe_Right - 400114: Toetip_Left - 400116: Toetip_Right - 400118: Trigger_Right - 400120: UpperArm_Left - 400122: UpperArm_Right - 400124: UpperLeg_Left - 400126: UpperLeg_Right - 2100000: w_h_01 - 2100002: w_b_01 - 2100004: w_bp_01 - 2100006: w_a_01 - 2100008: w_f_01 - 2100010: w_w_01 - 4300000: head1 - 4300002: Body1 - 4300004: Backpack1 - 4300006: Arm1 - 4300008: Leg1 - 4300010: AssaultRifle - 7400000: Idle_Ducking_AR - 9500000: //RootNode - 13700000: Arm1 - 13700002: AssaultRifle - 13700004: Backpack1 - 13700006: Body1 - 13700008: head1 - 13700010: Leg1 - 2186277476908879412: ImportLogs + serializedVersion: 19301 + internalIDToNameTable: + - first: + 1: 100000 + second: Arm1 + - first: + 1: 100002 + second: ArmPlacement_Left + - first: + 1: 100004 + second: ArmPlacement_Right + - first: + 1: 100006 + second: ArmPlacement_Upper + - first: + 1: 100008 + second: ArmPosition_Left + - first: + 1: 100010 + second: ArmPosition_Right + - first: + 1: 100012 + second: AssaultRifle + - first: + 1: 100014 + second: BackPack + - first: + 1: 100016 + second: Backpack1 + - first: + 1: 100018 + second: Body1 + - first: + 1: 100020 + second: Chest + - first: + 1: 100022 + second: Foot_Left + - first: + 1: 100024 + second: Foot_Right + - first: + 1: 100026 + second: Hand_Left + - first: + 1: 100028 + second: Hand_Right + - first: + 1: 100030 + second: Head + - first: + 1: 100032 + second: head1 + - first: + 1: 100034 + second: Headgear_Left + - first: + 1: 100036 + second: Headgear_Right + - first: + 1: 100038 + second: Hips + - first: + 1: 100040 + second: //RootNode + - first: + 1: 100042 + second: Index_Distal_Left + - first: + 1: 100044 + second: Index_Distal_Right + - first: + 1: 100046 + second: Index_Intermediate_Left + - first: + 1: 100048 + second: Index_Intermediate_Right + - first: + 1: 100050 + second: Index_Proximal_Left + - first: + 1: 100052 + second: Index_Proximal_Right + - first: + 1: 100054 + second: Leg1 + - first: + 1: 100056 + second: LowerArm_Left + - first: + 1: 100058 + second: LowerArm_Right + - first: + 1: 100060 + second: LowerLeg_Left + - first: + 1: 100062 + second: LowerLeg_Right + - first: + 1: 100064 + second: magazine_Right + - first: + 1: 100066 + second: Neck + - first: + 1: 100068 + second: RestOfFingers_Distal_Left + - first: + 1: 100070 + second: RestOfFingers_Distal_Right + - first: + 1: 100072 + second: RestOfFingers_Intermediate_Left + - first: + 1: 100074 + second: RestOfFingers_Intermediate_Right + - first: + 1: 100076 + second: RestOfFingers_Proximal_Left + - first: + 1: 100078 + second: RestOfFingers_Proximal_Right + - first: + 1: 100080 + second: Shoulder_Left + - first: + 1: 100082 + second: Shoulder_Right + - first: + 1: 100084 + second: ShoulderPadBlade_Left + - first: + 1: 100086 + second: ShoulderPadBlade_Right + - first: + 1: 100088 + second: ShoulderPadBody_Left + - first: + 1: 100090 + second: ShoulderPadBody_Right + - first: + 1: 100092 + second: ShoulderPadCTRL_Left + - first: + 1: 100094 + second: ShoulderPadCTRL_Right + - first: + 1: 100096 + second: Spine + - first: + 1: 100098 + second: Thumb_Distal_Left + - first: + 1: 100100 + second: Thumb_Distal_Right + - first: + 1: 100102 + second: Thumb_Intermediate_Left + - first: + 1: 100104 + second: Thumb_Intermediate_Right + - first: + 1: 100106 + second: Thumb_Proximal_Left + - first: + 1: 100108 + second: Thumb_Proximal_Right + - first: + 1: 100110 + second: Toe_Left + - first: + 1: 100112 + second: Toe_Right + - first: + 1: 100114 + second: Toetip_Left + - first: + 1: 100116 + second: Toetip_Right + - first: + 1: 100118 + second: Trigger_Right + - first: + 1: 100120 + second: UpperArm_Left + - first: + 1: 100122 + second: UpperArm_Right + - first: + 1: 100124 + second: UpperLeg_Left + - first: + 1: 100126 + second: UpperLeg_Right + - first: + 4: 400000 + second: Arm1 + - first: + 4: 400002 + second: ArmPlacement_Left + - first: + 4: 400004 + second: ArmPlacement_Right + - first: + 4: 400006 + second: ArmPlacement_Upper + - first: + 4: 400008 + second: ArmPosition_Left + - first: + 4: 400010 + second: ArmPosition_Right + - first: + 4: 400012 + second: AssaultRifle + - first: + 4: 400014 + second: BackPack + - first: + 4: 400016 + second: Backpack1 + - first: + 4: 400018 + second: Body1 + - first: + 4: 400020 + second: Chest + - first: + 4: 400022 + second: Foot_Left + - first: + 4: 400024 + second: Foot_Right + - first: + 4: 400026 + second: Hand_Left + - first: + 4: 400028 + second: Hand_Right + - first: + 4: 400030 + second: Head + - first: + 4: 400032 + second: head1 + - first: + 4: 400034 + second: Headgear_Left + - first: + 4: 400036 + second: Headgear_Right + - first: + 4: 400038 + second: Hips + - first: + 4: 400040 + second: //RootNode + - first: + 4: 400042 + second: Index_Distal_Left + - first: + 4: 400044 + second: Index_Distal_Right + - first: + 4: 400046 + second: Index_Intermediate_Left + - first: + 4: 400048 + second: Index_Intermediate_Right + - first: + 4: 400050 + second: Index_Proximal_Left + - first: + 4: 400052 + second: Index_Proximal_Right + - first: + 4: 400054 + second: Leg1 + - first: + 4: 400056 + second: LowerArm_Left + - first: + 4: 400058 + second: LowerArm_Right + - first: + 4: 400060 + second: LowerLeg_Left + - first: + 4: 400062 + second: LowerLeg_Right + - first: + 4: 400064 + second: magazine_Right + - first: + 4: 400066 + second: Neck + - first: + 4: 400068 + second: RestOfFingers_Distal_Left + - first: + 4: 400070 + second: RestOfFingers_Distal_Right + - first: + 4: 400072 + second: RestOfFingers_Intermediate_Left + - first: + 4: 400074 + second: RestOfFingers_Intermediate_Right + - first: + 4: 400076 + second: RestOfFingers_Proximal_Left + - first: + 4: 400078 + second: RestOfFingers_Proximal_Right + - first: + 4: 400080 + second: Shoulder_Left + - first: + 4: 400082 + second: Shoulder_Right + - first: + 4: 400084 + second: ShoulderPadBlade_Left + - first: + 4: 400086 + second: ShoulderPadBlade_Right + - first: + 4: 400088 + second: ShoulderPadBody_Left + - first: + 4: 400090 + second: ShoulderPadBody_Right + - first: + 4: 400092 + second: ShoulderPadCTRL_Left + - first: + 4: 400094 + second: ShoulderPadCTRL_Right + - first: + 4: 400096 + second: Spine + - first: + 4: 400098 + second: Thumb_Distal_Left + - first: + 4: 400100 + second: Thumb_Distal_Right + - first: + 4: 400102 + second: Thumb_Intermediate_Left + - first: + 4: 400104 + second: Thumb_Intermediate_Right + - first: + 4: 400106 + second: Thumb_Proximal_Left + - first: + 4: 400108 + second: Thumb_Proximal_Right + - first: + 4: 400110 + second: Toe_Left + - first: + 4: 400112 + second: Toe_Right + - first: + 4: 400114 + second: Toetip_Left + - first: + 4: 400116 + second: Toetip_Right + - first: + 4: 400118 + second: Trigger_Right + - first: + 4: 400120 + second: UpperArm_Left + - first: + 4: 400122 + second: UpperArm_Right + - first: + 4: 400124 + second: UpperLeg_Left + - first: + 4: 400126 + second: UpperLeg_Right + - first: + 21: 2100000 + second: w_h_01 + - first: + 21: 2100002 + second: w_b_01 + - first: + 21: 2100004 + second: w_bp_01 + - first: + 21: 2100006 + second: w_a_01 + - first: + 21: 2100008 + second: w_f_01 + - first: + 21: 2100010 + second: w_w_01 + - first: + 43: 4300000 + second: head1 + - first: + 43: 4300002 + second: Body1 + - first: + 43: 4300004 + second: Backpack1 + - first: + 43: 4300006 + second: Arm1 + - first: + 43: 4300008 + second: Leg1 + - first: + 43: 4300010 + second: AssaultRifle + - first: + 74: 7400000 + second: Idle_Ducking_AR + - first: + 95: 9500000 + second: //RootNode + - first: + 137: 13700000 + second: Arm1 + - first: + 137: 13700002 + second: AssaultRifle + - first: + 137: 13700004 + second: Backpack1 + - first: + 137: 13700006 + second: Body1 + - first: + 137: 13700008 + second: head1 + - first: + 137: 13700010 + second: Leg1 + - first: + 41386430: 2186277476908879412 + second: ImportLogs externalObjects: {} materials: - importMaterials: 1 + materialImportMode: 1 materialName: 0 materialSearch: 1 materialLocation: 1 @@ -183,6 +481,7 @@ ModelImporter: - serializedVersion: 16 name: Idle_Ducking_AR takeName: Take 001 + internalID: 0 firstFrame: 0 lastFrame: 30 wrapMode: 0 @@ -191,12 +490,12 @@ ModelImporter: cycleOffset: 0 loop: 0 hasAdditiveReferencePose: 0 - loopTime: 1 + loopTime: 0 loopBlend: 0 loopBlendOrientation: 1 loopBlendPositionY: 1 loopBlendPositionXZ: 1 - keepOriginalOrientation: 1 + keepOriginalOrientation: 0 keepOriginalPositionY: 1 keepOriginalPositionXZ: 1 heightFromFeet: 0 @@ -343,25 +642,28 @@ ModelImporter: meshCompression: 0 addColliders: 0 useSRGBMaterialColor: 1 + sortHierarchyByName: 1 importVisibility: 1 importBlendShapes: 1 importCameras: 1 importLights: 1 + fileIdsGeneration: 1 swapUVChannels: 0 generateSecondaryUV: 0 useFileUnits: 1 - optimizeMeshForGPU: 1 keepQuads: 0 weldVertices: 1 preserveHierarchy: 0 + skinWeightsMode: 0 + maxBonesPerVertex: 4 + minBoneWeight: 0.001 + meshOptimizationFlags: -1 indexFormat: 0 secondaryUVAngleDistortion: 8 secondaryUVAreaDistortion: 15.000001 secondaryUVHardAngle: 88 secondaryUVPackMargin: 4 useFileScale: 1 - previousCalculatedGlobalScale: 1 - hasPreviousCalculatedGlobalScale: 1 tangentSpace: normalSmoothAngle: 60 normalImportMode: 0 @@ -370,10 +672,10 @@ ModelImporter: legacyComputeAllNormalsFromSmoothingGroupsWhenMeshHasBlendShapes: 0 blendShapeNormalImportMode: 1 normalSmoothingSource: 0 + referencedClips: [] importAnimation: 1 - copyAvatar: 0 humanDescription: - serializedVersion: 2 + serializedVersion: 3 human: - boneName: Hips humanName: Hips @@ -1015,13 +1317,16 @@ ModelImporter: armStretch: 0.05 legStretch: 0.05 feetSpacing: 0 + globalScale: 1 rootMotionBoneName: hasTranslationDoF: 0 hasExtraRoot: 1 skeletonHasParents: 1 lastHumanDescriptionAvatarSource: {instanceID: 0} + autoGenerateAvatarMappingIfUnspecified: 1 animationType: 3 humanoidOversampling: 1 + avatarSetup: 1 additionalBone: 0 userData: assetBundleName: diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_Shoot_ar.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_guard_ar.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx.meta old mode 100755 new mode 100644 similarity index 79% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx.meta index b7fa658..61e2bcd --- a/Assets/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx.meta +++ b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Idle_gunMiddle_ar.fbx.meta @@ -1,160 +1,458 @@ fileFormatVersion: 2 guid: 107649059ea401b4e9c5c20f21e99a55 ModelImporter: - serializedVersion: 23 - fileIDToRecycleName: - 100000: Arm1 - 100002: ArmPlacement_Left - 100004: ArmPlacement_Right - 100006: ArmPlacement_Upper - 100008: ArmPosition_Left - 100010: ArmPosition_Right - 100012: AssaultRifle - 100014: BackPack - 100016: Backpack1 - 100018: Body1 - 100020: Chest - 100022: Foot_Left - 100024: Foot_Right - 100026: Hand_Left - 100028: Hand_Right - 100030: Head - 100032: head1 - 100034: Headgear_Left - 100036: Headgear_Right - 100038: Hips - 100040: //RootNode - 100042: Index_Distal_Left - 100044: Index_Distal_Right - 100046: Index_Intermediate_Left - 100048: Index_Intermediate_Right - 100050: Index_Proximal_Left - 100052: Index_Proximal_Right - 100054: Leg1 - 100056: LowerArm_Left - 100058: LowerArm_Right - 100060: LowerLeg_Left - 100062: LowerLeg_Right - 100064: magazine_Right - 100066: Neck - 100068: RestOfFingers_Distal_Left - 100070: RestOfFingers_Distal_Right - 100072: RestOfFingers_Intermediate_Left - 100074: RestOfFingers_Intermediate_Right - 100076: RestOfFingers_Proximal_Left - 100078: RestOfFingers_Proximal_Right - 100080: Shoulder_Left - 100082: Shoulder_Right - 100084: ShoulderPadBlade_Left - 100086: ShoulderPadBlade_Right - 100088: ShoulderPadBody_Left - 100090: ShoulderPadBody_Right - 100092: ShoulderPadCTRL_Left - 100094: ShoulderPadCTRL_Right - 100096: Spine - 100098: Thumb_Distal_Left - 100100: Thumb_Distal_Right - 100102: Thumb_Intermediate_Left - 100104: Thumb_Intermediate_Right - 100106: Thumb_Proximal_Left - 100108: Thumb_Proximal_Right - 100110: Toe_Left - 100112: Toe_Right - 100114: Toetip_Left - 100116: Toetip_Right - 100118: Trigger_Right - 100120: UpperArm_Left - 100122: UpperArm_Right - 100124: UpperLeg_Left - 100126: UpperLeg_Right - 400000: Arm1 - 400002: ArmPlacement_Left - 400004: ArmPlacement_Right - 400006: ArmPlacement_Upper - 400008: ArmPosition_Left - 400010: ArmPosition_Right - 400012: AssaultRifle - 400014: BackPack - 400016: Backpack1 - 400018: Body1 - 400020: Chest - 400022: Foot_Left - 400024: Foot_Right - 400026: Hand_Left - 400028: Hand_Right - 400030: Head - 400032: head1 - 400034: Headgear_Left - 400036: Headgear_Right - 400038: Hips - 400040: //RootNode - 400042: Index_Distal_Left - 400044: Index_Distal_Right - 400046: Index_Intermediate_Left - 400048: Index_Intermediate_Right - 400050: Index_Proximal_Left - 400052: Index_Proximal_Right - 400054: Leg1 - 400056: LowerArm_Left - 400058: LowerArm_Right - 400060: LowerLeg_Left - 400062: LowerLeg_Right - 400064: magazine_Right - 400066: Neck - 400068: RestOfFingers_Distal_Left - 400070: RestOfFingers_Distal_Right - 400072: RestOfFingers_Intermediate_Left - 400074: RestOfFingers_Intermediate_Right - 400076: RestOfFingers_Proximal_Left - 400078: RestOfFingers_Proximal_Right - 400080: Shoulder_Left - 400082: Shoulder_Right - 400084: ShoulderPadBlade_Left - 400086: ShoulderPadBlade_Right - 400088: ShoulderPadBody_Left - 400090: ShoulderPadBody_Right - 400092: ShoulderPadCTRL_Left - 400094: ShoulderPadCTRL_Right - 400096: Spine - 400098: Thumb_Distal_Left - 400100: Thumb_Distal_Right - 400102: Thumb_Intermediate_Left - 400104: Thumb_Intermediate_Right - 400106: Thumb_Proximal_Left - 400108: Thumb_Proximal_Right - 400110: Toe_Left - 400112: Toe_Right - 400114: Toetip_Left - 400116: Toetip_Right - 400118: Trigger_Right - 400120: UpperArm_Left - 400122: UpperArm_Right - 400124: UpperLeg_Left - 400126: UpperLeg_Right - 2100000: w_h_01 - 2100002: w_b_01 - 2100004: w_bp_01 - 2100006: w_a_01 - 2100008: w_f_01 - 2100010: w_w_01 - 4300000: head1 - 4300002: Body1 - 4300004: Backpack1 - 4300006: Arm1 - 4300008: Leg1 - 4300010: AssaultRifle - 7400000: Idle_gunMiddle_AR - 9500000: //RootNode - 13700000: Arm1 - 13700002: AssaultRifle - 13700004: Backpack1 - 13700006: Body1 - 13700008: head1 - 13700010: Leg1 - 2186277476908879412: ImportLogs + serializedVersion: 19301 + internalIDToNameTable: + - first: + 1: 100000 + second: Arm1 + - first: + 1: 100002 + second: ArmPlacement_Left + - first: + 1: 100004 + second: ArmPlacement_Right + - first: + 1: 100006 + second: ArmPlacement_Upper + - first: + 1: 100008 + second: ArmPosition_Left + - first: + 1: 100010 + second: ArmPosition_Right + - first: + 1: 100012 + second: AssaultRifle + - first: + 1: 100014 + second: BackPack + - first: + 1: 100016 + second: Backpack1 + - first: + 1: 100018 + second: Body1 + - first: + 1: 100020 + second: Chest + - first: + 1: 100022 + second: Foot_Left + - first: + 1: 100024 + second: Foot_Right + - first: + 1: 100026 + second: Hand_Left + - first: + 1: 100028 + second: Hand_Right + - first: + 1: 100030 + second: Head + - first: + 1: 100032 + second: head1 + - first: + 1: 100034 + second: Headgear_Left + - first: + 1: 100036 + second: Headgear_Right + - first: + 1: 100038 + second: Hips + - first: + 1: 100040 + second: //RootNode + - first: + 1: 100042 + second: Index_Distal_Left + - first: + 1: 100044 + second: Index_Distal_Right + - first: + 1: 100046 + second: Index_Intermediate_Left + - first: + 1: 100048 + second: Index_Intermediate_Right + - first: + 1: 100050 + second: Index_Proximal_Left + - first: + 1: 100052 + second: Index_Proximal_Right + - first: + 1: 100054 + second: Leg1 + - first: + 1: 100056 + second: LowerArm_Left + - first: + 1: 100058 + second: LowerArm_Right + - first: + 1: 100060 + second: LowerLeg_Left + - first: + 1: 100062 + second: LowerLeg_Right + - first: + 1: 100064 + second: magazine_Right + - first: + 1: 100066 + second: Neck + - first: + 1: 100068 + second: RestOfFingers_Distal_Left + - first: + 1: 100070 + second: RestOfFingers_Distal_Right + - first: + 1: 100072 + second: RestOfFingers_Intermediate_Left + - first: + 1: 100074 + second: RestOfFingers_Intermediate_Right + - first: + 1: 100076 + second: RestOfFingers_Proximal_Left + - first: + 1: 100078 + second: RestOfFingers_Proximal_Right + - first: + 1: 100080 + second: Shoulder_Left + - first: + 1: 100082 + second: Shoulder_Right + - first: + 1: 100084 + second: ShoulderPadBlade_Left + - first: + 1: 100086 + second: ShoulderPadBlade_Right + - first: + 1: 100088 + second: ShoulderPadBody_Left + - first: + 1: 100090 + second: ShoulderPadBody_Right + - first: + 1: 100092 + second: ShoulderPadCTRL_Left + - first: + 1: 100094 + second: ShoulderPadCTRL_Right + - first: + 1: 100096 + second: Spine + - first: + 1: 100098 + second: Thumb_Distal_Left + - first: + 1: 100100 + second: Thumb_Distal_Right + - first: + 1: 100102 + second: Thumb_Intermediate_Left + - first: + 1: 100104 + second: Thumb_Intermediate_Right + - first: + 1: 100106 + second: Thumb_Proximal_Left + - first: + 1: 100108 + second: Thumb_Proximal_Right + - first: + 1: 100110 + second: Toe_Left + - first: + 1: 100112 + second: Toe_Right + - first: + 1: 100114 + second: Toetip_Left + - first: + 1: 100116 + second: Toetip_Right + - first: + 1: 100118 + second: Trigger_Right + - first: + 1: 100120 + second: UpperArm_Left + - first: + 1: 100122 + second: UpperArm_Right + - first: + 1: 100124 + second: UpperLeg_Left + - first: + 1: 100126 + second: UpperLeg_Right + - first: + 4: 400000 + second: Arm1 + - first: + 4: 400002 + second: ArmPlacement_Left + - first: + 4: 400004 + second: ArmPlacement_Right + - first: + 4: 400006 + second: ArmPlacement_Upper + - first: + 4: 400008 + second: ArmPosition_Left + - first: + 4: 400010 + second: ArmPosition_Right + - first: + 4: 400012 + second: AssaultRifle + - first: + 4: 400014 + second: BackPack + - first: + 4: 400016 + second: Backpack1 + - first: + 4: 400018 + second: Body1 + - first: + 4: 400020 + second: Chest + - first: + 4: 400022 + second: Foot_Left + - first: + 4: 400024 + second: Foot_Right + - first: + 4: 400026 + second: Hand_Left + - first: + 4: 400028 + second: Hand_Right + - first: + 4: 400030 + second: Head + - first: + 4: 400032 + second: head1 + - first: + 4: 400034 + second: Headgear_Left + - first: + 4: 400036 + second: Headgear_Right + - first: + 4: 400038 + second: Hips + - first: + 4: 400040 + second: //RootNode + - first: + 4: 400042 + second: Index_Distal_Left + - first: + 4: 400044 + second: Index_Distal_Right + - first: + 4: 400046 + second: Index_Intermediate_Left + - first: + 4: 400048 + second: Index_Intermediate_Right + - first: + 4: 400050 + second: Index_Proximal_Left + - first: + 4: 400052 + second: Index_Proximal_Right + - first: + 4: 400054 + second: Leg1 + - first: + 4: 400056 + second: LowerArm_Left + - first: + 4: 400058 + second: LowerArm_Right + - first: + 4: 400060 + second: LowerLeg_Left + - first: + 4: 400062 + second: LowerLeg_Right + - first: + 4: 400064 + second: magazine_Right + - first: + 4: 400066 + second: Neck + - first: + 4: 400068 + second: RestOfFingers_Distal_Left + - first: + 4: 400070 + second: RestOfFingers_Distal_Right + - first: + 4: 400072 + second: RestOfFingers_Intermediate_Left + - first: + 4: 400074 + second: RestOfFingers_Intermediate_Right + - first: + 4: 400076 + second: RestOfFingers_Proximal_Left + - first: + 4: 400078 + second: RestOfFingers_Proximal_Right + - first: + 4: 400080 + second: Shoulder_Left + - first: + 4: 400082 + second: Shoulder_Right + - first: + 4: 400084 + second: ShoulderPadBlade_Left + - first: + 4: 400086 + second: ShoulderPadBlade_Right + - first: + 4: 400088 + second: ShoulderPadBody_Left + - first: + 4: 400090 + second: ShoulderPadBody_Right + - first: + 4: 400092 + second: ShoulderPadCTRL_Left + - first: + 4: 400094 + second: ShoulderPadCTRL_Right + - first: + 4: 400096 + second: Spine + - first: + 4: 400098 + second: Thumb_Distal_Left + - first: + 4: 400100 + second: Thumb_Distal_Right + - first: + 4: 400102 + second: Thumb_Intermediate_Left + - first: + 4: 400104 + second: Thumb_Intermediate_Right + - first: + 4: 400106 + second: Thumb_Proximal_Left + - first: + 4: 400108 + second: Thumb_Proximal_Right + - first: + 4: 400110 + second: Toe_Left + - first: + 4: 400112 + second: Toe_Right + - first: + 4: 400114 + second: Toetip_Left + - first: + 4: 400116 + second: Toetip_Right + - first: + 4: 400118 + second: Trigger_Right + - first: + 4: 400120 + second: UpperArm_Left + - first: + 4: 400122 + second: UpperArm_Right + - first: + 4: 400124 + second: UpperLeg_Left + - first: + 4: 400126 + second: UpperLeg_Right + - first: + 21: 2100000 + second: w_h_01 + - first: + 21: 2100002 + second: w_b_01 + - first: + 21: 2100004 + second: w_bp_01 + - first: + 21: 2100006 + second: w_a_01 + - first: + 21: 2100008 + second: w_f_01 + - first: + 21: 2100010 + second: w_w_01 + - first: + 43: 4300000 + second: head1 + - first: + 43: 4300002 + second: Body1 + - first: + 43: 4300004 + second: Backpack1 + - first: + 43: 4300006 + second: Arm1 + - first: + 43: 4300008 + second: Leg1 + - first: + 43: 4300010 + second: AssaultRifle + - first: + 74: 7400000 + second: Idle_gunMiddle_AR + - first: + 95: 9500000 + second: //RootNode + - first: + 137: 13700000 + second: Arm1 + - first: + 137: 13700002 + second: AssaultRifle + - first: + 137: 13700004 + second: Backpack1 + - first: + 137: 13700006 + second: Body1 + - first: + 137: 13700008 + second: head1 + - first: + 137: 13700010 + second: Leg1 + - first: + 41386430: 2186277476908879412 + second: ImportLogs externalObjects: {} materials: - importMaterials: 1 + materialImportMode: 1 materialName: 0 materialSearch: 1 materialLocation: 1 @@ -183,6 +481,7 @@ ModelImporter: - serializedVersion: 16 name: Idle_gunMiddle_AR takeName: Take 001 + internalID: 0 firstFrame: 0 lastFrame: 40 wrapMode: 0 @@ -343,25 +642,28 @@ ModelImporter: meshCompression: 0 addColliders: 0 useSRGBMaterialColor: 1 + sortHierarchyByName: 1 importVisibility: 1 importBlendShapes: 1 importCameras: 1 importLights: 1 + fileIdsGeneration: 1 swapUVChannels: 0 generateSecondaryUV: 0 useFileUnits: 1 - optimizeMeshForGPU: 1 keepQuads: 0 weldVertices: 1 preserveHierarchy: 0 + skinWeightsMode: 0 + maxBonesPerVertex: 4 + minBoneWeight: 0.001 + meshOptimizationFlags: -1 indexFormat: 0 secondaryUVAngleDistortion: 8 secondaryUVAreaDistortion: 15.000001 secondaryUVHardAngle: 88 secondaryUVPackMargin: 4 useFileScale: 1 - previousCalculatedGlobalScale: 1 - hasPreviousCalculatedGlobalScale: 1 tangentSpace: normalSmoothAngle: 60 normalImportMode: 0 @@ -370,10 +672,10 @@ ModelImporter: legacyComputeAllNormalsFromSmoothingGroupsWhenMeshHasBlendShapes: 0 blendShapeNormalImportMode: 1 normalSmoothingSource: 0 + referencedClips: [] importAnimation: 1 - copyAvatar: 0 humanDescription: - serializedVersion: 2 + serializedVersion: 3 human: - boneName: Hips humanName: Hips @@ -1015,13 +1317,16 @@ ModelImporter: armStretch: 0.05 legStretch: 0.05 feetSpacing: 0 + globalScale: 1 rootMotionBoneName: hasTranslationDoF: 0 hasExtraRoot: 1 skeletonHasParents: 1 lastHumanDescriptionAvatarSource: {instanceID: 0} + autoGenerateAvatarMappingIfUnspecified: 1 animationType: 3 humanoidOversampling: 1 + avatarSetup: 1 additionalBone: 0 userData: assetBundleName: diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Jump.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Reload.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_guard_AR.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Run_gunMiddle_AR.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_AutoShot_AR.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_BurstShot_AR.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/Shoot_SingleShot_AR.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkBack_Shoot_ar.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkFront_Shoot_ar.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkLeft_Shoot_ar.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animations/WalkRight_Shoot_ar.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animators.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animators.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animators.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animators.meta diff --git a/Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller new file mode 100644 index 0000000..01ed567 --- /dev/null +++ b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller @@ -0,0 +1,93 @@ +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!206 &-2926378548289824311 +BlendTree: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Blend Tree + m_Childs: [] + m_BlendParameter: Blend + m_BlendParameterY: Blend + m_MinThreshold: 0 + m_MaxThreshold: 1 + m_UseAutomaticThresholds: 1 + m_NormalizedBlendValues: 0 + m_BlendType: 0 +--- !u!91 &9100000 +AnimatorController: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: SciFiWarrior + serializedVersion: 5 + m_AnimatorParameters: + - m_Name: Blend + m_Type: 1 + m_DefaultFloat: 0 + m_DefaultInt: 0 + m_DefaultBool: 0 + m_Controller: {fileID: 0} + m_AnimatorLayers: + - serializedVersion: 5 + m_Name: Base Layer + m_StateMachine: {fileID: 1107476691846109156} + m_Mask: {fileID: 0} + m_Motions: [] + m_Behaviours: [] + m_BlendingMode: 0 + m_SyncedLayerIndex: -1 + m_DefaultWeight: 0 + m_IKPass: 0 + m_SyncedLayerAffectsTiming: 0 + m_Controller: {fileID: 9100000} +--- !u!1107 &1107476691846109156 +AnimatorStateMachine: + serializedVersion: 6 + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Base Layer + m_ChildStates: + - serializedVersion: 1 + m_State: {fileID: 5405818589905925533} + m_Position: {x: 353.98297, y: 140.66443, z: 0} + m_ChildStateMachines: [] + m_AnyStateTransitions: [] + m_EntryTransitions: [] + m_StateMachineTransitions: {} + m_StateMachineBehaviours: [] + m_AnyStatePosition: {x: 50, y: 20, z: 0} + m_EntryPosition: {x: 50, y: 120, z: 0} + m_ExitPosition: {x: 48, y: 72, z: 0} + m_ParentStateMachinePosition: {x: 800, y: 20, z: 0} + m_DefaultState: {fileID: 5405818589905925533} +--- !u!1102 &5405818589905925533 +AnimatorState: + serializedVersion: 6 + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Blend Tree + m_Speed: 1 + m_CycleOffset: 0 + m_Transitions: [] + m_StateMachineBehaviours: [] + m_Position: {x: 50, y: 50, z: 0} + m_IKOnFeet: 0 + m_WriteDefaultValues: 1 + m_Mirror: 0 + m_SpeedParameterActive: 0 + m_MirrorParameterActive: 0 + m_CycleOffsetParameterActive: 0 + m_TimeParameterActive: 0 + m_Motion: {fileID: 7400000, guid: 107649059ea401b4e9c5c20f21e99a55, type: 3} + m_Tag: + m_SpeedParameter: + m_MirrorParameter: + m_CycleOffsetParameter: + m_TimeParameter: diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Animators/SciFiWarriorOur.controller.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/HP.mat b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/HP.mat old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/HP.mat rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/HP.mat diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/HP.mat.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/HP.mat.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/HP.mat.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/HP.mat.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/PBR.mat b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/PBR.mat old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/PBR.mat rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/PBR.mat diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/PBR.mat.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/PBR.mat.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/PBR.mat.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/PBR.mat.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Polyart.mat.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat old mode 100755 new mode 100644 similarity index 96% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat index a1934e5..ba73d5d --- a/Assets/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat +++ b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat @@ -1,89 +1,89 @@ -%YAML 1.1 -%TAG !u! tag:unity3d.com,2011: ---- !u!21 &2100000 -Material: - serializedVersion: 6 - m_ObjectHideFlags: 0 - m_PrefabParentObject: {fileID: 0} - m_PrefabInternal: {fileID: 0} - m_Name: Skybox_Mat - m_Shader: {fileID: 106, guid: 0000000000000000f000000000000000, type: 0} - m_ShaderKeywords: _METALLIC_SETUP _SUNDISK_HIGH_QUALITY - m_LightmapFlags: 4 - m_EnableInstancingVariants: 0 - m_DoubleSidedGI: 0 - m_CustomRenderQueue: -1 - stringTagMap: {} - disabledShaderPasses: [] - m_SavedProperties: - serializedVersion: 3 - m_TexEnvs: - - _BumpMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _DetailAlbedoMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _DetailMask: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _DetailNormalMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _EmissionMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _MainTex: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _MetallicGlossMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _OcclusionMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _ParallaxMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _SpecGlossMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - m_Floats: - - _AtmosphereThickness: 0.53 - - _BumpScale: 1 - - _Cutoff: 0.5 - - _DetailNormalMapScale: 1 - - _DstBlend: 0 - - _Exposure: 1.25 - - _GlossMapScale: 1 - - _Glossiness: 0.5 - - _GlossyReflections: 1 - - _Metallic: 0 - - _Mode: 0 - - _OcclusionStrength: 1 - - _Parallax: 0.02 - - _SmoothnessTextureChannel: 0 - - _SpecularHighlights: 1 - - _SrcBlend: 1 - - _SunDisk: 2 - - _SunSize: 0.04 - - _SunSizeConvergence: 5 - - _UVSec: 0 - - _WorkflowMode: 1 - - _ZWrite: 1 - m_Colors: - - _Color: {r: 1, g: 1, b: 1, a: 1} - - _EmissionColor: {r: 0, g: 0, b: 0, a: 1} - - _GroundColor: {r: 0.6392157, g: 0.6901961, b: 0.7411765, a: 1} - - _SkyTint: {r: 1, g: 1, b: 1, a: 1} - - _SpecColor: {r: 0.19999996, g: 0.19999996, b: 0.19999996, a: 1} +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!21 &2100000 +Material: + serializedVersion: 6 + m_ObjectHideFlags: 0 + m_PrefabParentObject: {fileID: 0} + m_PrefabInternal: {fileID: 0} + m_Name: Skybox_Mat + m_Shader: {fileID: 106, guid: 0000000000000000f000000000000000, type: 0} + m_ShaderKeywords: _METALLIC_SETUP _SUNDISK_HIGH_QUALITY + m_LightmapFlags: 4 + m_EnableInstancingVariants: 0 + m_DoubleSidedGI: 0 + m_CustomRenderQueue: -1 + stringTagMap: {} + disabledShaderPasses: [] + m_SavedProperties: + serializedVersion: 3 + m_TexEnvs: + - _BumpMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _DetailAlbedoMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _DetailMask: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _DetailNormalMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _EmissionMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _MainTex: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _MetallicGlossMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _OcclusionMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _ParallaxMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + - _SpecGlossMap: + m_Texture: {fileID: 0} + m_Scale: {x: 1, y: 1} + m_Offset: {x: 0, y: 0} + m_Floats: + - _AtmosphereThickness: 0.53 + - _BumpScale: 1 + - _Cutoff: 0.5 + - _DetailNormalMapScale: 1 + - _DstBlend: 0 + - _Exposure: 1.25 + - _GlossMapScale: 1 + - _Glossiness: 0.5 + - _GlossyReflections: 1 + - _Metallic: 0 + - _Mode: 0 + - _OcclusionStrength: 1 + - _Parallax: 0.02 + - _SmoothnessTextureChannel: 0 + - _SpecularHighlights: 1 + - _SrcBlend: 1 + - _SunDisk: 2 + - _SunSize: 0.04 + - _SunSizeConvergence: 5 + - _UVSec: 0 + - _WorkflowMode: 1 + - _ZWrite: 1 + m_Colors: + - _Color: {r: 1, g: 1, b: 1, a: 1} + - _EmissionColor: {r: 0, g: 0, b: 0, a: 1} + - _GroundColor: {r: 0.6392157, g: 0.6901961, b: 0.7411765, a: 1} + - _SkyTint: {r: 1, g: 1, b: 1, a: 1} + - _SpecColor: {r: 0.19999996, g: 0.19999996, b: 0.19999996, a: 1} diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Skybox_Mat.mat.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/Stage.mat b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Stage.mat old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/Stage.mat rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Stage.mat diff --git a/Assets/SciFiWarriorPBRHPPolyart/Materials/Stage.mat.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Stage.mat.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Materials/Stage.mat.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Materials/Stage.mat.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Mesh.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Mesh.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/PBR_HP_Mesh.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Polyart_Mesh.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx b/Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx diff --git a/Assets/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Mesh/Stage.fbx.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Prefabs.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Prefabs.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab diff --git a/Assets/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/HPCharacter.prefab.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab diff --git a/Assets/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PBRCharacter.prefab.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab old mode 100755 new mode 100644 similarity index 96% rename from Assets/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab index 67bf0a4..56bc88c --- a/Assets/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab +++ b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab @@ -1715,6 +1715,10 @@ GameObject: m_Component: - component: {fileID: 7072444061591351794} - component: {fileID: 7072444061582351952} + - component: {fileID: 6381261399840543104} + - component: {fileID: 2971219656579870553} + - component: {fileID: 5912052353413311719} + - component: {fileID: 8293113947138725737} m_Layer: 0 m_Name: PolyartCharacter m_TagString: Untagged @@ -1762,6 +1766,104 @@ Animator: m_HasTransformHierarchy: 1 m_AllowConstantClipSamplingOptimization: 1 m_KeepAnimatorControllerStateOnDisable: 0 +--- !u!136 &6381261399840543104 +CapsuleCollider: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444061591711186} + m_Material: {fileID: 0} + m_IsTrigger: 0 + m_Enabled: 1 + m_Radius: 0.5 + m_Height: 2.1 + m_Direction: 1 + m_Center: {x: 0, y: 1.06, z: 0} +--- !u!114 &2971219656579870553 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444061591711186} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 9826297ef4d853741b2af768441ec7f7, type: 3} + m_Name: + m_EditorClassIdentifier: + input_View: {x: 0, y: 0} + cameraHolder: {fileID: 7072444061591351696} + feetTransform: {fileID: 7072444061591351794} + playerSettings: + ViewXSensetivity: 0 + ViewYSensetivity: 0 + ViewXInverted: 0 + ViewYInverted: 0 + SprintingHold: 0 + MovementSmoothing: 0 + RunningForwardSpeed: 0 + RunningStrafeSpeed: 0 + WalkingForwardSpeed: 0 + WalkingBackwardSpeed: 0 + WalkingStrafeSpeed: 0 + JumpingHeight: 0 + JumpingFalloff: 0 + FallingSmoothing: 0 + SpeedEffector: 1 + CrouchSpeedEffector: 0 + ProneSpeedEffector: 0 + FallingSpeedEffector: 0 + ViewClampYMin: -70 + ViewClampYMax: 80 + playerMask: + serializedVersion: 2 + m_Bits: 0 + gravityAmount: 9.81 + gravityMin: 0 + jumpingForce: {x: 0, y: 30, z: 0} + playerStance: 0 + playerStanceSmoothing: 0 + playerStandStance: + CameraHeight: 0 + StanceCollider: {fileID: 6381261399840543104} + playerCrouchStance: + CameraHeight: 0 + StanceCollider: {fileID: 0} + playerProneStance: + CameraHeight: 0 + StanceCollider: {fileID: 0} + currentWeapon: {fileID: 0} +--- !u!114 &5912052353413311719 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444061591711186} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 907ff02de47a55a4e971d73d25e7d006, type: 3} + m_Name: + m_EditorClassIdentifier: +--- !u!143 &8293113947138725737 +CharacterController: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444061591711186} + m_Material: {fileID: 0} + m_IsTrigger: 0 + m_Enabled: 1 + serializedVersion: 2 + m_Height: 2.1 + m_Radius: 0.5 + m_SlopeLimit: 45 + m_StepOffset: 0.3 + m_SkinWidth: 0.08 + m_MinMoveDistance: 0.001 + m_Center: {x: 0, y: 1.02, z: 0} --- !u!1 &7072444061591711188 GameObject: m_ObjectHideFlags: 0 diff --git a/Assets/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Prefabs/PolyartCharacter.prefab.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Scene.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Scene.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Scene.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Scene.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity b/Assets/Arts/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity diff --git a/Assets/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Scene/DemoScene01.unity.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/HP_Albedo.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Albedo.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_AO.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_EM.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_MS.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PBR_Free_NM.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/Polyart.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/Polyart.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/Polyart.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/Polyart.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/Polyart.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/Polyart.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/Polyart.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/Polyart.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png diff --git a/Assets/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/Textures/PolyartEmission.png.meta diff --git a/Assets/SciFiWarriorPBRHPPolyart/promotion.jpg b/Assets/Arts/SciFiWarriorPBRHPPolyart/promotion.jpg old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/promotion.jpg rename to Assets/Arts/SciFiWarriorPBRHPPolyart/promotion.jpg diff --git a/Assets/SciFiWarriorPBRHPPolyart/promotion.jpg.meta b/Assets/Arts/SciFiWarriorPBRHPPolyart/promotion.jpg.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/SciFiWarriorPBRHPPolyart/promotion.jpg.meta rename to Assets/Arts/SciFiWarriorPBRHPPolyart/promotion.jpg.meta diff --git a/Assets/Defaults/mat_proBuilder.mat b/Assets/Defaults/mat_proBuilder.mat deleted file mode 100755 index 0e86a32..0000000 --- a/Assets/Defaults/mat_proBuilder.mat +++ /dev/null @@ -1,77 +0,0 @@ -%YAML 1.1 -%TAG !u! tag:unity3d.com,2011: ---- !u!21 &2100000 -Material: - serializedVersion: 6 - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: mat_proBuilder - m_Shader: {fileID: 46, guid: 0000000000000000f000000000000000, type: 0} - m_ShaderKeywords: - m_LightmapFlags: 4 - m_EnableInstancingVariants: 0 - m_DoubleSidedGI: 0 - m_CustomRenderQueue: -1 - stringTagMap: {} - disabledShaderPasses: [] - m_SavedProperties: - serializedVersion: 3 - m_TexEnvs: - - _BumpMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _DetailAlbedoMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _DetailMask: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _DetailNormalMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _EmissionMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _MainTex: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _MetallicGlossMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _OcclusionMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - - _ParallaxMap: - m_Texture: {fileID: 0} - m_Scale: {x: 1, y: 1} - m_Offset: {x: 0, y: 0} - m_Floats: - - _BumpScale: 1 - - _Cutoff: 0.5 - - _DetailNormalMapScale: 1 - - _DstBlend: 0 - - _GlossMapScale: 1 - - _Glossiness: 0.5 - - _GlossyReflections: 1 - - _Metallic: 0 - - _Mode: 0 - - _OcclusionStrength: 1 - - _Parallax: 0.02 - - _SmoothnessTextureChannel: 0 - - _SpecularHighlights: 1 - - _SrcBlend: 1 - - _UVSec: 0 - - _ZWrite: 1 - m_Colors: - - _Color: {r: 0.735849, g: 0.13536847, b: 0.13536847, a: 1} - - _EmissionColor: {r: 0, g: 0, b: 0, a: 1} diff --git a/Assets/Inputs/InputSystem.inputsettings.asset b/Assets/Inputs/InputSystem.inputsettings.asset new file mode 100644 index 0000000..8a242b0 --- /dev/null +++ b/Assets/Inputs/InputSystem.inputsettings.asset @@ -0,0 +1,35 @@ +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!114 &11400000 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 0} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: c46f07b5ed07e4e92aa78254188d3d10, type: 3} + m_Name: InputSystem.inputsettings + m_EditorClassIdentifier: + m_SupportedDevices: [] + m_UpdateMode: 1 + m_MaxEventBytesPerUpdate: 5242880 + m_MaxQueuedEventsPerUpdate: 1000 + m_CompensateForScreenOrientation: 1 + m_BackgroundBehavior: 0 + m_EditorInputBehaviorInPlayMode: 0 + m_DefaultDeadzoneMin: 0.125 + m_DefaultDeadzoneMax: 0.925 + m_DefaultButtonPressPoint: 0.5 + m_ButtonReleaseThreshold: 0.75 + m_DefaultTapTime: 0.2 + m_DefaultSlowTapTime: 0.5 + m_DefaultHoldTime: 0.4 + m_TapRadius: 5 + m_MultiTapDelayTime: 0.75 + m_DisableRedundantEventsMerging: 0 + m_iOSSettings: + m_MotionUsage: + m_Enabled: 0 + m_Description: diff --git a/Assets/Inputs/InputSystem.inputsettings.asset.meta b/Assets/Inputs/InputSystem.inputsettings.asset.meta new file mode 100644 index 0000000..20c5cb0 --- /dev/null +++ b/Assets/Inputs/InputSystem.inputsettings.asset.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: 4480bcbc35319b14588f3c0eb33e88c8 +NativeFormatImporter: + externalObjects: {} + mainObjectFileID: 11400000 + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Inputs/ThirdPersonViewInput.cs b/Assets/Inputs/ThirdPersonViewInput.cs new file mode 100644 index 0000000..89ed85f --- /dev/null +++ b/Assets/Inputs/ThirdPersonViewInput.cs @@ -0,0 +1,354 @@ +//------------------------------------------------------------------------------ +// +// This code was auto-generated by com.unity.inputsystem:InputActionCodeGenerator +// version 1.3.0 +// from Assets/Inputs/ThirdPersonViewInput.inputactions +// +// Changes to this file may cause incorrect behavior and will be lost if +// the code is regenerated. +// +//------------------------------------------------------------------------------ + +using System; +using System.Collections; +using System.Collections.Generic; +using UnityEngine.InputSystem; +using UnityEngine.InputSystem.Utilities; + +public partial class @ThirdPersonViewInput : IInputActionCollection2, IDisposable +{ + public InputActionAsset asset { get; } + public @ThirdPersonViewInput() + { + asset = InputActionAsset.FromJson(@"{ + ""name"": ""ThirdPersonViewInput"", + ""maps"": [ + { + ""name"": ""Player Movement"", + ""id"": ""50a259c9-d17f-4483-bd4b-feb4e290989d"", + ""actions"": [ + { + ""name"": ""Movement"", + ""type"": ""PassThrough"", + ""id"": ""be8e85db-0110-45cd-823f-1058ed7eb184"", + ""expectedControlType"": ""Vector2"", + ""processors"": """", + ""interactions"": """", + ""initialStateCheck"": false + }, + { + ""name"": ""Camera"", + ""type"": ""PassThrough"", + ""id"": ""f7ad7ce9-d43f-468d-ab0f-54cc66ac33fa"", + ""expectedControlType"": ""Vector2"", + ""processors"": """", + ""interactions"": """", + ""initialStateCheck"": false + } + ], + ""bindings"": [ + { + ""name"": ""WASD"", + ""id"": ""72e37b69-1a0d-4e6e-a5aa-7a2a62845d22"", + ""path"": ""2DVector(mode=2)"", + ""interactions"": """", + ""processors"": """", + ""groups"": """", + ""action"": ""Movement"", + ""isComposite"": true, + ""isPartOfComposite"": false + }, + { + ""name"": ""up"", + ""id"": ""a5e60c10-53e0-484a-be45-210fc2d3d305"", + ""path"": ""/w"", + ""interactions"": """", + ""processors"": """", + ""groups"": """", + ""action"": ""Movement"", + ""isComposite"": false, + ""isPartOfComposite"": true + }, + { + ""name"": ""down"", + ""id"": ""cdafb043-b1bc-4680-9d55-91854d1e2eb4"", + ""path"": ""/s"", + ""interactions"": """", + ""processors"": """", + ""groups"": """", + ""action"": ""Movement"", + ""isComposite"": false, + ""isPartOfComposite"": true + }, + { + ""name"": ""left"", + ""id"": ""4da8623e-6baa-4256-99ec-d016f9ad6c1f"", + ""path"": ""/a"", + ""interactions"": """", + ""processors"": """", + ""groups"": """", + ""action"": ""Movement"", + ""isComposite"": false, + ""isPartOfComposite"": true + }, + { + ""name"": ""right"", + ""id"": ""5c98d073-e749-49de-bb73-c6f5920f53f1"", + ""path"": ""/d"", + ""interactions"": """", + ""processors"": """", + ""groups"": """", + ""action"": ""Movement"", + ""isComposite"": false, + ""isPartOfComposite"": true + }, + { + ""name"": """", + ""id"": ""d8db2334-67bd-461e-92dc-dc04f2b64660"", + ""path"": ""/delta"", + ""interactions"": """", + ""processors"": ""NormalizeVector2"", + ""groups"": """", + ""action"": ""Camera"", + ""isComposite"": false, + ""isPartOfComposite"": false + } + ] + }, + { + ""name"": ""Player Actions"", + ""id"": ""7c28b681-44ce-4c44-832c-561051fceb49"", + ""actions"": [ + { + ""name"": ""Jump"", + ""type"": ""Button"", + ""id"": ""ef60df41-7d24-45f4-a426-4e6b9fe4d374"", + ""expectedControlType"": ""Button"", + ""processors"": """", + ""interactions"": """", + ""initialStateCheck"": false + }, + { + ""name"": ""Crouch"", + ""type"": ""Button"", + ""id"": ""19075beb-f2d7-4af5-bbcc-e2b53e1e139a"", + ""expectedControlType"": ""Button"", + ""processors"": """", + ""interactions"": """", + ""initialStateCheck"": false + }, + { + ""name"": ""Fire"", + ""type"": ""PassThrough"", + ""id"": ""2b493869-f7ab-4acb-8918-11b0265f0993"", + ""expectedControlType"": ""Button"", + ""processors"": """", + ""interactions"": """", + ""initialStateCheck"": false + } + ], + ""bindings"": [ + { + ""name"": """", + ""id"": ""e8b4d9c1-7565-4954-b6f9-f9c7ed95f613"", + ""path"": ""/space"", + ""interactions"": """", + ""processors"": """", + ""groups"": """", + ""action"": ""Jump"", + ""isComposite"": false, + ""isPartOfComposite"": false + }, + { + ""name"": """", + ""id"": ""a02f0962-5e78-44c2-aaf3-ee059376bda6"", + ""path"": ""/c"", + ""interactions"": """", + ""processors"": """", + ""groups"": """", + ""action"": ""Crouch"", + ""isComposite"": false, + ""isPartOfComposite"": false + }, + { + ""name"": """", + ""id"": ""5caa47c3-a8d5-4aad-86fe-40cdca87210a"", + ""path"": ""/leftButton"", + ""interactions"": ""Hold(duration=0.1)"", + ""processors"": """", + ""groups"": """", + ""action"": ""Fire"", + ""isComposite"": false, + ""isPartOfComposite"": false + } + ] + } + ], + ""controlSchemes"": [] +}"); + // Player Movement + m_PlayerMovement = asset.FindActionMap("Player Movement", throwIfNotFound: true); + m_PlayerMovement_Movement = m_PlayerMovement.FindAction("Movement", throwIfNotFound: true); + m_PlayerMovement_Camera = m_PlayerMovement.FindAction("Camera", throwIfNotFound: true); + // Player Actions + m_PlayerActions = asset.FindActionMap("Player Actions", throwIfNotFound: true); + m_PlayerActions_Jump = m_PlayerActions.FindAction("Jump", throwIfNotFound: true); + m_PlayerActions_Crouch = m_PlayerActions.FindAction("Crouch", throwIfNotFound: true); + m_PlayerActions_Fire = m_PlayerActions.FindAction("Fire", throwIfNotFound: true); + } + + public void Dispose() + { + UnityEngine.Object.Destroy(asset); + } + + public InputBinding? bindingMask + { + get => asset.bindingMask; + set => asset.bindingMask = value; + } + + public ReadOnlyArray? devices + { + get => asset.devices; + set => asset.devices = value; + } + + public ReadOnlyArray controlSchemes => asset.controlSchemes; + + public bool Contains(InputAction action) + { + return asset.Contains(action); + } + + public IEnumerator GetEnumerator() + { + return asset.GetEnumerator(); + } + + IEnumerator IEnumerable.GetEnumerator() + { + return GetEnumerator(); + } + + public void Enable() + { + asset.Enable(); + } + + public void Disable() + { + asset.Disable(); + } + public IEnumerable bindings => asset.bindings; + + public InputAction FindAction(string actionNameOrId, bool throwIfNotFound = false) + { + return asset.FindAction(actionNameOrId, throwIfNotFound); + } + public int FindBinding(InputBinding bindingMask, out InputAction action) + { + return asset.FindBinding(bindingMask, out action); + } + + // Player Movement + private readonly InputActionMap m_PlayerMovement; + private IPlayerMovementActions m_PlayerMovementActionsCallbackInterface; + private readonly InputAction m_PlayerMovement_Movement; + private readonly InputAction m_PlayerMovement_Camera; + public struct PlayerMovementActions + { + private @ThirdPersonViewInput m_Wrapper; + public PlayerMovementActions(@ThirdPersonViewInput wrapper) { m_Wrapper = wrapper; } + public InputAction @Movement => m_Wrapper.m_PlayerMovement_Movement; + public InputAction @Camera => m_Wrapper.m_PlayerMovement_Camera; + public InputActionMap Get() { return m_Wrapper.m_PlayerMovement; } + public void Enable() { Get().Enable(); } + public void Disable() { Get().Disable(); } + public bool enabled => Get().enabled; + public static implicit operator InputActionMap(PlayerMovementActions set) { return set.Get(); } + public void SetCallbacks(IPlayerMovementActions instance) + { + if (m_Wrapper.m_PlayerMovementActionsCallbackInterface != null) + { + @Movement.started -= m_Wrapper.m_PlayerMovementActionsCallbackInterface.OnMovement; + @Movement.performed -= m_Wrapper.m_PlayerMovementActionsCallbackInterface.OnMovement; + @Movement.canceled -= m_Wrapper.m_PlayerMovementActionsCallbackInterface.OnMovement; + @Camera.started -= m_Wrapper.m_PlayerMovementActionsCallbackInterface.OnCamera; + @Camera.performed -= m_Wrapper.m_PlayerMovementActionsCallbackInterface.OnCamera; + @Camera.canceled -= m_Wrapper.m_PlayerMovementActionsCallbackInterface.OnCamera; + } + m_Wrapper.m_PlayerMovementActionsCallbackInterface = instance; + if (instance != null) + { + @Movement.started += instance.OnMovement; + @Movement.performed += instance.OnMovement; + @Movement.canceled += instance.OnMovement; + @Camera.started += instance.OnCamera; + @Camera.performed += instance.OnCamera; + @Camera.canceled += instance.OnCamera; + } + } + } + public PlayerMovementActions @PlayerMovement => new PlayerMovementActions(this); + + // Player Actions + private readonly InputActionMap m_PlayerActions; + private IPlayerActionsActions m_PlayerActionsActionsCallbackInterface; + private readonly InputAction m_PlayerActions_Jump; + private readonly InputAction m_PlayerActions_Crouch; + private readonly InputAction m_PlayerActions_Fire; + public struct PlayerActionsActions + { + private @ThirdPersonViewInput m_Wrapper; + public PlayerActionsActions(@ThirdPersonViewInput wrapper) { m_Wrapper = wrapper; } + public InputAction @Jump => m_Wrapper.m_PlayerActions_Jump; + public InputAction @Crouch => m_Wrapper.m_PlayerActions_Crouch; + public InputAction @Fire => m_Wrapper.m_PlayerActions_Fire; + public InputActionMap Get() { return m_Wrapper.m_PlayerActions; } + public void Enable() { Get().Enable(); } + public void Disable() { Get().Disable(); } + public bool enabled => Get().enabled; + public static implicit operator InputActionMap(PlayerActionsActions set) { return set.Get(); } + public void SetCallbacks(IPlayerActionsActions instance) + { + if (m_Wrapper.m_PlayerActionsActionsCallbackInterface != null) + { + @Jump.started -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnJump; + @Jump.performed -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnJump; + @Jump.canceled -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnJump; + @Crouch.started -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnCrouch; + @Crouch.performed -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnCrouch; + @Crouch.canceled -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnCrouch; + @Fire.started -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnFire; + @Fire.performed -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnFire; + @Fire.canceled -= m_Wrapper.m_PlayerActionsActionsCallbackInterface.OnFire; + } + m_Wrapper.m_PlayerActionsActionsCallbackInterface = instance; + if (instance != null) + { + @Jump.started += instance.OnJump; + @Jump.performed += instance.OnJump; + @Jump.canceled += instance.OnJump; + @Crouch.started += instance.OnCrouch; + @Crouch.performed += instance.OnCrouch; + @Crouch.canceled += instance.OnCrouch; + @Fire.started += instance.OnFire; + @Fire.performed += instance.OnFire; + @Fire.canceled += instance.OnFire; + } + } + } + public PlayerActionsActions @PlayerActions => new PlayerActionsActions(this); + public interface IPlayerMovementActions + { + void OnMovement(InputAction.CallbackContext context); + void OnCamera(InputAction.CallbackContext context); + } + public interface IPlayerActionsActions + { + void OnJump(InputAction.CallbackContext context); + void OnCrouch(InputAction.CallbackContext context); + void OnFire(InputAction.CallbackContext context); + } +} diff --git a/Assets/Inputs/ThirdPersonViewInput.cs.meta b/Assets/Inputs/ThirdPersonViewInput.cs.meta new file mode 100644 index 0000000..724326c --- /dev/null +++ b/Assets/Inputs/ThirdPersonViewInput.cs.meta @@ -0,0 +1,11 @@ +fileFormatVersion: 2 +guid: e6b0597c7a248fd4e8e5ca2ef5d7f29f +MonoImporter: + externalObjects: {} + serializedVersion: 2 + defaultReferences: [] + executionOrder: 0 + icon: {instanceID: 0} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Inputs/ThirdPersonViewInput.inputactions b/Assets/Inputs/ThirdPersonViewInput.inputactions new file mode 100644 index 0000000..a0c0388 --- /dev/null +++ b/Assets/Inputs/ThirdPersonViewInput.inputactions @@ -0,0 +1,166 @@ +{ + "name": "ThirdPersonViewInput", + "maps": [ + { + "name": "Player Movement", + "id": "50a259c9-d17f-4483-bd4b-feb4e290989d", + "actions": [ + { + "name": "Movement", + "type": "PassThrough", + "id": "be8e85db-0110-45cd-823f-1058ed7eb184", + "expectedControlType": "Vector2", + "processors": "", + "interactions": "", + "initialStateCheck": false + }, + { + "name": "Camera", + "type": "PassThrough", + "id": "f7ad7ce9-d43f-468d-ab0f-54cc66ac33fa", + "expectedControlType": "Vector2", + "processors": "", + "interactions": "", + "initialStateCheck": false + } + ], + "bindings": [ + { + "name": "WASD", + "id": "72e37b69-1a0d-4e6e-a5aa-7a2a62845d22", + "path": "2DVector(mode=2)", + "interactions": "", + "processors": "", + "groups": "", + "action": "Movement", + "isComposite": true, + "isPartOfComposite": false + }, + { + "name": "up", + "id": "a5e60c10-53e0-484a-be45-210fc2d3d305", + "path": "/w", + "interactions": "", + "processors": "", + "groups": "", + "action": "Movement", + "isComposite": false, + "isPartOfComposite": true + }, + { + "name": "down", + "id": "cdafb043-b1bc-4680-9d55-91854d1e2eb4", + "path": "/s", + "interactions": "", + "processors": "", + "groups": "", + "action": "Movement", + "isComposite": false, + "isPartOfComposite": true + }, + { + "name": "left", + "id": "4da8623e-6baa-4256-99ec-d016f9ad6c1f", + "path": "/a", + "interactions": "", + "processors": "", + "groups": "", + "action": "Movement", + "isComposite": false, + "isPartOfComposite": true + }, + { + "name": "right", + "id": "5c98d073-e749-49de-bb73-c6f5920f53f1", + "path": "/d", + "interactions": "", + "processors": "", + "groups": "", + "action": "Movement", + "isComposite": false, + "isPartOfComposite": true + }, + { + "name": "", + "id": "d8db2334-67bd-461e-92dc-dc04f2b64660", + "path": "/delta", + "interactions": "", + "processors": "NormalizeVector2", + "groups": "", + "action": "Camera", + "isComposite": false, + "isPartOfComposite": false + } + ] + }, + { + "name": "Player Actions", + "id": "7c28b681-44ce-4c44-832c-561051fceb49", + "actions": [ + { + "name": "Jump", + "type": "Button", + "id": "ef60df41-7d24-45f4-a426-4e6b9fe4d374", + "expectedControlType": "Button", + "processors": "", + "interactions": "", + "initialStateCheck": false + }, + { + "name": "Crouch", + "type": "Button", + "id": "19075beb-f2d7-4af5-bbcc-e2b53e1e139a", + "expectedControlType": "Button", + "processors": "", + "interactions": "", + "initialStateCheck": false + }, + { + "name": "Fire", + "type": "PassThrough", + "id": "2b493869-f7ab-4acb-8918-11b0265f0993", + "expectedControlType": "Button", + "processors": "", + "interactions": "", + "initialStateCheck": false + } + ], + "bindings": [ + { + "name": "", + "id": "e8b4d9c1-7565-4954-b6f9-f9c7ed95f613", + "path": "/space", + "interactions": "", + "processors": "", + "groups": "", + "action": "Jump", + "isComposite": false, + "isPartOfComposite": false + }, + { + "name": "", + "id": "a02f0962-5e78-44c2-aaf3-ee059376bda6", + "path": "/c", + "interactions": "", + "processors": "", + "groups": "", + "action": "Crouch", + "isComposite": false, + "isPartOfComposite": false + }, + { + "name": "", + "id": "5caa47c3-a8d5-4aad-86fe-40cdca87210a", + "path": "/leftButton", + "interactions": "Hold(duration=0.1)", + "processors": "", + "groups": "", + "action": "Fire", + "isComposite": false, + "isPartOfComposite": false + } + ] + } + ], + "controlSchemes": [] +} \ No newline at end of file diff --git a/Assets/Inputs/ThirdPersonViewInput.inputactions.meta b/Assets/Inputs/ThirdPersonViewInput.inputactions.meta new file mode 100644 index 0000000..2e8e4a7 --- /dev/null +++ b/Assets/Inputs/ThirdPersonViewInput.inputactions.meta @@ -0,0 +1,14 @@ +fileFormatVersion: 2 +guid: a17c8887bfca8834d8287fab2f0cadf8 +ScriptedImporter: + internalIDToNameTable: [] + externalObjects: {} + serializedVersion: 2 + userData: + assetBundleName: + assetBundleVariant: + script: {fileID: 11500000, guid: 8404be70184654265930450def6a9037, type: 3} + generateWrapperCode: 1 + wrapperCodePath: + wrapperClassName: + wrapperCodeNamespace: diff --git a/Assets/ML-Agents/Timers/DemoScene01_timers.json b/Assets/ML-Agents/Timers/DemoScene01_timers.json deleted file mode 100755 index 3914503..0000000 --- a/Assets/ML-Agents/Timers/DemoScene01_timers.json +++ /dev/null @@ -1 +0,0 @@ -{"count":1,"self":249.99626239999998,"total":250.70272609999998,"children":{"InitializeActuators":{"count":1,"self":0.0005131,"total":0.0005131,"children":null},"AgentSendState":{"count":8389,"self":0.1187579,"total":0.1187579,"children":null},"DecideAction":{"count":8389,"self":0.5871921,"total":0.5871921,"children":null}},"gauges":{},"metadata":{"timer_format_version":"0.1.0","start_time_seconds":"1649259643","unity_version":"2019.4.36f1","command_line_arguments":"C:\\Program Files\\Unity\\Hub\\Editor\\2019.4.36f1\\Editor\\Unity.exe -projectpath D:\\real_shooter -useHub -hubIPC -cloudEnvironment production -licensingIpc LicenseClient-user -hubSessionId 91406950-b5ad-11ec-a63c-e7b76cbae13d -accessToken EBt97pMhHqClFDnjD_uh-3vplxP-uI2yS0WK-hSxfuM012f","communication_protocol_version":"1.5.0","com.unity.ml-agents_version":"2.0.1","scene_name":"DemoScene01","end_time_seconds":"1649259893"}} \ No newline at end of file diff --git a/Assets/ML-Agents/Timers/dont touch me plz_timers.json b/Assets/ML-Agents/Timers/dont touch me plz_timers.json deleted file mode 100755 index 699e26c..0000000 --- a/Assets/ML-Agents/Timers/dont touch me plz_timers.json +++ /dev/null @@ -1 +0,0 @@ -{"count":1,"self":14.5494256,"total":14.5766034,"children":{"InitializeActuators":{"count":1,"self":0,"total":0,"children":null},"AgentSendState":{"count":497,"self":0.0020012999999999997,"total":0.0020012999999999997,"children":null},"DecideAction":{"count":497,"self":0.0241768,"total":0.0241768,"children":null}},"gauges":{},"metadata":{"timer_format_version":"0.1.0","start_time_seconds":"1649741696","unity_version":"2019.4.35f1","command_line_arguments":"D:\\Unity\\2019.4.35f1\\Editor\\Unity.exe -projectpath C:\\Users\\kiril\\real-shooter -useHub -hubIPC -cloudEnvironment production -licensingIpc LicenseClient-kiril -hubSessionId 21e9de90-ba14-11ec-9398-079edccf5e34 -accessToken oIpf_7gKWQQRilVTeJehUsFhrxasdzsG_K3j5Swtgx0009f","communication_protocol_version":"1.5.0","com.unity.ml-agents_version":"2.0.1","scene_name":"dont touch me plz","end_time_seconds":"1649741711"}} \ No newline at end of file diff --git a/Assets/Prefabs/CharacterPrefabs.meta b/Assets/Prefabs/CharacterPrefabs.meta new file mode 100644 index 0000000..2ca1002 --- /dev/null +++ b/Assets/Prefabs/CharacterPrefabs.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: a7a6fd6139b92ce43af2a2f1116ea410 +folderAsset: yes +DefaultImporter: + externalObjects: {} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Prefabs/Bot.prefab b/Assets/Prefabs/CharacterPrefabs/Bot.prefab similarity index 100% rename from Assets/Prefabs/Bot.prefab rename to Assets/Prefabs/CharacterPrefabs/Bot.prefab diff --git a/Assets/Prefabs/Bot.prefab.meta b/Assets/Prefabs/CharacterPrefabs/Bot.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/Bot.prefab.meta rename to Assets/Prefabs/CharacterPrefabs/Bot.prefab.meta diff --git a/Assets/Prefabs/CharacterPrefabs/Character.prefab b/Assets/Prefabs/CharacterPrefabs/Character.prefab new file mode 100644 index 0000000..0699e0c --- /dev/null +++ b/Assets/Prefabs/CharacterPrefabs/Character.prefab @@ -0,0 +1,2639 @@ +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!1 &6633355367928763904 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859616} + m_Layer: 0 + m_Name: Thumb_Intermediate_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859616 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763904} + m_LocalRotation: {x: -0.04900198, y: -0.13549508, z: 0.10049181, w: 0.98444974} + m_LocalPosition: {x: -9.835772, y: -1.1368684e-13, z: -2.842171e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859620} + m_Father: {fileID: 6633355367928859628} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763906 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859618} + m_Layer: 0 + m_Name: Thumb_Distal_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859618 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763906} + m_LocalRotation: {x: 0.0017341404, y: 0.008320127, z: -0.005799853, w: 0.9999471} + m_LocalPosition: {x: -8.816269, y: 0, z: -1.4210855e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859630} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763908 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859620} + m_Layer: 0 + m_Name: Thumb_Distal_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859620 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763908} + m_LocalRotation: {x: -0.017883137, y: -0.03974522, z: -0.025908269, w: 0.99871385} + m_LocalPosition: {x: -8.816608, y: 0, z: 1.4210855e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859616} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763910 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859622} + m_Layer: 0 + m_Name: Spine + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859622 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763910} + m_LocalRotation: {x: 0.99452215, y: -0.104526356, z: -7.7878193e-10, w: 0.0000000074097675} + m_LocalPosition: {x: -21.090727, y: -8.881784e-15, z: -4.4013775e-16} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859538} + m_Father: {fileID: 6633355367928859552} + m_RootOrder: 2 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763912 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859624} + m_Layer: 0 + m_Name: Toe_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859624 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763912} + m_LocalRotation: {x: 0.0013659683, y: 0.0051824837, z: -0.24785845, w: 0.9687815} + m_LocalPosition: {x: -13.771131, y: 1.9539925e-14, z: -1.0658141e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859636} + m_Father: {fileID: 6633355367928859536} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763914 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859626} + m_Layer: 0 + m_Name: Thumb_Proximal_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859626 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763914} + m_LocalRotation: {x: -0.057529792, y: -0.49499094, z: 0.11207554, w: 0.859717} + m_LocalPosition: {x: -0.29218963, y: -0.56312394, z: -12.29586} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859630} + m_Father: {fileID: 6633355367928859546} + m_RootOrder: 2 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763916 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859628} + m_Layer: 0 + m_Name: Thumb_Proximal_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859628 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763916} + m_LocalRotation: {x: 0.057522308, y: 0.49499637, z: 0.112048574, w: 0.8597179} + m_LocalPosition: {x: -0.2922163, y: -0.5636321, z: 12.295864} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859616} + m_Father: {fileID: 6633355367928859548} + m_RootOrder: 2 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763918 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859630} + m_Layer: 0 + m_Name: Thumb_Intermediate_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859630 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763918} + m_LocalRotation: {x: 0.048999686, y: 0.1354733, z: 0.1004816, w: 0.9844539} + m_LocalPosition: {x: -9.836, y: -5.684342e-14, z: -5.684342e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859618} + m_Father: {fileID: 6633355367928859626} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763920 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859632} + m_Layer: 0 + m_Name: Trigger_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859632 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763920} + m_LocalRotation: {x: -0.0031178175, y: 0.0071248533, z: 0.40088162, w: 0.91609687} + m_LocalPosition: {x: -9.913989, y: -10.731702, z: 9.284221} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859532} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763922 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859634} + m_Layer: 0 + m_Name: Toetip_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859634 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763922} + m_LocalRotation: {x: -0.00028878966, y: 0.016934738, z: -0.012053749, w: 0.99978393} + m_LocalPosition: {x: -17.883856, y: -2.6645353e-15, z: -1.4210855e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859638} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763924 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859636} + m_Layer: 0 + m_Name: Toetip_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859636 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763924} + m_LocalRotation: {x: 0.00016529544, y: -0.016936114, z: -0.0047631934, w: 0.99984527} + m_LocalPosition: {x: -17.883871, y: 3.5527137e-15, z: 2.1316282e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859624} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763926 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859638} + m_Layer: 0 + m_Name: Toe_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859638 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763926} + m_LocalRotation: {x: -0.0013655907, y: -0.005181019, z: -0.24785995, w: 0.9687811} + m_LocalPosition: {x: -13.771146, y: 2.1316282e-14, z: 1.0658141e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859634} + m_Father: {fileID: 6633355367928859550} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763928 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859640} + m_Layer: 0 + m_Name: UpperLeg_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859640 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763928} + m_LocalRotation: {x: 0.03061261, y: 0.6954094, z: 0.7139454, w: 0.07583304} + m_LocalPosition: {x: 0.084786385, y: 0.40212917, z: 18.8642} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859578} + m_Father: {fileID: 6633355367928859552} + m_RootOrder: 4 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763930 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859642} + m_Layer: 0 + m_Name: UpperLeg_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859642 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763930} + m_LocalRotation: {x: -0.030611672, y: -0.6954005, z: 0.7139541, w: 0.07583354} + m_LocalPosition: {x: 0.08523814, y: 0.40205857, z: -18.864191} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859580} + m_Father: {fileID: 6633355367928859552} + m_RootOrder: 3 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763932 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859644} + m_Layer: 0 + m_Name: UpperArm_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859644 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763932} + m_LocalRotation: {x: 0.029697837, y: 0.15896995, z: -0.18122736, w: 0.97005326} + m_LocalPosition: {x: -15.111769, y: 0, z: -3.1974423e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859582} + m_Father: {fileID: 6633355367928859604} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928763934 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859646} + m_Layer: 0 + m_Name: UpperArm_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859646 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928763934} + m_LocalRotation: {x: -0.029700447, y: -0.15896967, z: -0.18123563, w: 0.97005165} + m_LocalPosition: {x: -15.111847, y: 0, z: -3.7303494e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859568} + m_Father: {fileID: 6633355367928859606} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764320 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859520} + m_Layer: 0 + m_Name: ArmPlacement_Upper + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859520 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764320} + m_LocalRotation: {x: -0.002378591, y: 0.08787313, z: 0.02695381, w: 0.99576414} + m_LocalPosition: {x: -0.56803536, y: 46.34834, z: 10.906936} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859528} + m_RootOrder: 2 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764322 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859522} + m_Layer: 0 + m_Name: ArmPlacement_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859522 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764322} + m_LocalRotation: {x: -0.02695381, y: 0.99576414, z: -0.0023785909, w: -0.08787313} + m_LocalPosition: {x: 41.471436, y: 4.6548405, z: -15.306103} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859528} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764324 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859524} + m_Layer: 0 + m_Name: ArmPlacement_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859524 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764324} + m_LocalRotation: {x: -0.002378591, y: 0.08787313, z: 0.02695381, w: 0.99576414} + m_LocalPosition: {x: -44.450283, y: 2.842171e-14, z: -2.1316282e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859528} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764326 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859526} + - component: {fileID: 6633355367931943590} + m_Layer: 0 + m_Name: Arm1 + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859526 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764326} + m_LocalRotation: {x: 0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: -0, y: -163.22968, z: -0.3527179} + m_LocalScale: {x: 0.01, y: 0.01, z: 0.01} + m_Children: [] + m_Father: {fileID: 6633355367928859588} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!137 &6633355367931943590 +SkinnedMeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764326} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 0 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 2100000, guid: 78660c09ff101ea4cb77de408a9527da, type: 2} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 0 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 + serializedVersion: 2 + m_Quality: 0 + m_UpdateWhenOffscreen: 0 + m_SkinnedMotionVectors: 1 + m_Mesh: {fileID: 4300006, guid: 860793eb3324391468f1c120a75ec049, type: 3} + m_Bones: + - {fileID: 6633355367928859552} + - {fileID: 6633355367928859622} + - {fileID: 6633355367928859538} + - {fileID: 6633355367928859590} + - {fileID: 6633355367928859544} + - {fileID: 6633355367928859556} + - {fileID: 6633355367928859554} + - {fileID: 6633355367928859606} + - {fileID: 6633355367928859646} + - {fileID: 6633355367928859568} + - {fileID: 6633355367928859548} + - {fileID: 6633355367928859594} + - {fileID: 6633355367928859598} + - {fileID: 6633355367928859586} + - {fileID: 6633355367928859574} + - {fileID: 6633355367928859562} + - {fileID: 6633355367928859566} + - {fileID: 6633355367928859628} + - {fileID: 6633355367928859616} + - {fileID: 6633355367928859620} + - {fileID: 6633355367928859610} + - {fileID: 6633355367928859614} + - {fileID: 6633355367928859602} + - {fileID: 6633355367928859604} + - {fileID: 6633355367928859644} + - {fileID: 6633355367928859582} + - {fileID: 6633355367928859546} + - {fileID: 6633355367928859592} + - {fileID: 6633355367928859596} + - {fileID: 6633355367928859584} + - {fileID: 6633355367928859572} + - {fileID: 6633355367928859560} + - {fileID: 6633355367928859564} + - {fileID: 6633355367928859626} + - {fileID: 6633355367928859630} + - {fileID: 6633355367928859618} + - {fileID: 6633355367928859608} + - {fileID: 6633355367928859612} + - {fileID: 6633355367928859600} + - {fileID: 6633355367928859528} + - {fileID: 6633355367928859524} + - {fileID: 6633355367928859520} + - {fileID: 6633355367928859522} + - {fileID: 6633355367928859642} + - {fileID: 6633355367928859580} + - {fileID: 6633355367928859536} + - {fileID: 6633355367928859624} + - {fileID: 6633355367928859636} + - {fileID: 6633355367928859640} + - {fileID: 6633355367928859578} + - {fileID: 6633355367928859550} + - {fileID: 6633355367928859638} + - {fileID: 6633355367928859634} + - {fileID: 6633355367928859532} + - {fileID: 6633355367928859576} + - {fileID: 6633355367928859632} + - {fileID: 6633355367928859534} + m_BlendShapeWeights: [] + m_RootBone: {fileID: 6633355367928859552} + m_AABB: + m_Center: {x: -58.594677, y: 13.632675, z: 0.00010681152} + m_Extent: {x: 29.436275, y: 28.906914, z: 140.37485} + m_DirtyAABB: 0 +--- !u!1 &6633355367928764328 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859528} + m_Layer: 0 + m_Name: BackPack + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859528 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764328} + m_LocalRotation: {x: -0.3617453, y: 0.5306367, z: 0.3968756, w: 0.6557856} + m_LocalPosition: {x: -10.219411, y: -38.523506, z: 0.00000010329652} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859524} + - {fileID: 6633355367928859522} + - {fileID: 6633355367928859520} + m_Father: {fileID: 6633355367928859538} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764330 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859530} + - component: {fileID: 6633355367931943588} + m_Layer: 0 + m_Name: AssaultRifle + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859530 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764330} + m_LocalRotation: {x: 0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: -118.131454, y: -98.01, z: -57.969975} + m_LocalScale: {x: 0.01, y: 0.01, z: 0.01} + m_Children: [] + m_Father: {fileID: 6633355367928859588} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!137 &6633355367931943588 +SkinnedMeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764330} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 0 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 2100000, guid: 78660c09ff101ea4cb77de408a9527da, type: 2} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 0 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 + serializedVersion: 2 + m_Quality: 0 + m_UpdateWhenOffscreen: 0 + m_SkinnedMotionVectors: 1 + m_Mesh: {fileID: 4300010, guid: 860793eb3324391468f1c120a75ec049, type: 3} + m_Bones: + - {fileID: 6633355367928859552} + - {fileID: 6633355367928859622} + - {fileID: 6633355367928859538} + - {fileID: 6633355367928859590} + - {fileID: 6633355367928859544} + - {fileID: 6633355367928859556} + - {fileID: 6633355367928859554} + - {fileID: 6633355367928859606} + - {fileID: 6633355367928859646} + - {fileID: 6633355367928859568} + - {fileID: 6633355367928859548} + - {fileID: 6633355367928859594} + - {fileID: 6633355367928859598} + - {fileID: 6633355367928859586} + - {fileID: 6633355367928859574} + - {fileID: 6633355367928859562} + - {fileID: 6633355367928859566} + - {fileID: 6633355367928859628} + - {fileID: 6633355367928859616} + - {fileID: 6633355367928859620} + - {fileID: 6633355367928859610} + - {fileID: 6633355367928859614} + - {fileID: 6633355367928859602} + - {fileID: 6633355367928859604} + - {fileID: 6633355367928859644} + - {fileID: 6633355367928859582} + - {fileID: 6633355367928859546} + - {fileID: 6633355367928859592} + - {fileID: 6633355367928859596} + - {fileID: 6633355367928859584} + - {fileID: 6633355367928859572} + - {fileID: 6633355367928859560} + - {fileID: 6633355367928859564} + - {fileID: 6633355367928859626} + - {fileID: 6633355367928859630} + - {fileID: 6633355367928859618} + - {fileID: 6633355367928859608} + - {fileID: 6633355367928859612} + - {fileID: 6633355367928859600} + - {fileID: 6633355367928859528} + - {fileID: 6633355367928859524} + - {fileID: 6633355367928859520} + - {fileID: 6633355367928859522} + - {fileID: 6633355367928859642} + - {fileID: 6633355367928859580} + - {fileID: 6633355367928859536} + - {fileID: 6633355367928859624} + - {fileID: 6633355367928859636} + - {fileID: 6633355367928859640} + - {fileID: 6633355367928859578} + - {fileID: 6633355367928859550} + - {fileID: 6633355367928859638} + - {fileID: 6633355367928859634} + - {fileID: 6633355367928859532} + - {fileID: 6633355367928859576} + - {fileID: 6633355367928859632} + - {fileID: 6633355367928859534} + m_BlendShapeWeights: [] + m_RootBone: {fileID: 6633355367928859552} + m_AABB: + m_Center: {x: -34.27955, y: -49.079704, z: 120.48372} + m_Extent: {x: 95.48148, y: 94.6697, z: 10.629513} + m_DirtyAABB: 0 +--- !u!1 &6633355367928764332 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859532} + m_Layer: 0 + m_Name: ArmPosition_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859532 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764332} + m_LocalRotation: {x: 0.0025357183, y: -0.0073521743, z: -0.47313836, w: 0.88095385} + m_LocalPosition: {x: -2.423428, y: -55.614994, z: 120.53} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859576} + - {fileID: 6633355367928859632} + m_Father: {fileID: 6633355367928859552} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764334 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859534} + m_Layer: 0 + m_Name: ArmPosition_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859534 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764334} + m_LocalRotation: {x: 0.0073521743, y: 0.002535718, z: 0.88095385, w: 0.47313833} + m_LocalPosition: {x: -2.423428, y: -55.614994, z: -120.53} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859552} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764336 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859536} + m_Layer: 0 + m_Name: Foot_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859536 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764336} + m_LocalRotation: {x: 0.8193084, y: 0.56578285, z: -0.02233458, w: 0.09013736} + m_LocalPosition: {x: -62.734695, y: 2.842171e-14, z: -1.7763568e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859624} + m_Father: {fileID: 6633355367928859580} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764338 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859538} + m_Layer: 0 + m_Name: Chest + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859538 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764338} + m_LocalRotation: {x: 3.1378165e-33, y: -5.551116e-17, z: 0.18862787, w: 0.9820487} + m_LocalPosition: {x: -24.042006, y: -1.0658141e-14, z: 1.9451509e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859528} + - {fileID: 6633355367928859590} + - {fileID: 6633355367928859606} + - {fileID: 6633355367928859604} + m_Father: {fileID: 6633355367928859622} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764340 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859540} + - component: {fileID: 6633355367931943584} + m_Layer: 0 + m_Name: Body1 + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859540 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764340} + m_LocalRotation: {x: 0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: -0.0000018693923, y: -129.49721, z: -6.0212374} + m_LocalScale: {x: 0.01, y: 0.01, z: 0.01} + m_Children: [] + m_Father: {fileID: 6633355367928859588} + m_RootOrder: 3 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!137 &6633355367931943584 +SkinnedMeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764340} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 0 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 2100000, guid: 78660c09ff101ea4cb77de408a9527da, type: 2} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 0 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 + serializedVersion: 2 + m_Quality: 0 + m_UpdateWhenOffscreen: 0 + m_SkinnedMotionVectors: 1 + m_Mesh: {fileID: 4300002, guid: 860793eb3324391468f1c120a75ec049, type: 3} + m_Bones: + - {fileID: 6633355367928859552} + - {fileID: 6633355367928859622} + - {fileID: 6633355367928859538} + - {fileID: 6633355367928859590} + - {fileID: 6633355367928859544} + - {fileID: 6633355367928859556} + - {fileID: 6633355367928859554} + - {fileID: 6633355367928859606} + - {fileID: 6633355367928859646} + - {fileID: 6633355367928859568} + - {fileID: 6633355367928859548} + - {fileID: 6633355367928859594} + - {fileID: 6633355367928859598} + - {fileID: 6633355367928859586} + - {fileID: 6633355367928859574} + - {fileID: 6633355367928859562} + - {fileID: 6633355367928859566} + - {fileID: 6633355367928859628} + - {fileID: 6633355367928859616} + - {fileID: 6633355367928859620} + - {fileID: 6633355367928859610} + - {fileID: 6633355367928859614} + - {fileID: 6633355367928859602} + - {fileID: 6633355367928859604} + - {fileID: 6633355367928859644} + - {fileID: 6633355367928859582} + - {fileID: 6633355367928859546} + - {fileID: 6633355367928859592} + - {fileID: 6633355367928859596} + - {fileID: 6633355367928859584} + - {fileID: 6633355367928859572} + - {fileID: 6633355367928859560} + - {fileID: 6633355367928859564} + - {fileID: 6633355367928859626} + - {fileID: 6633355367928859630} + - {fileID: 6633355367928859618} + - {fileID: 6633355367928859608} + - {fileID: 6633355367928859612} + - {fileID: 6633355367928859600} + - {fileID: 6633355367928859528} + - {fileID: 6633355367928859524} + - {fileID: 6633355367928859520} + - {fileID: 6633355367928859522} + - {fileID: 6633355367928859642} + - {fileID: 6633355367928859580} + - {fileID: 6633355367928859536} + - {fileID: 6633355367928859624} + - {fileID: 6633355367928859636} + - {fileID: 6633355367928859640} + - {fileID: 6633355367928859578} + - {fileID: 6633355367928859550} + - {fileID: 6633355367928859638} + - {fileID: 6633355367928859634} + - {fileID: 6633355367928859532} + - {fileID: 6633355367928859576} + - {fileID: 6633355367928859632} + - {fileID: 6633355367928859534} + m_BlendShapeWeights: [] + m_RootBone: {fileID: 6633355367928859552} + m_AABB: + m_Center: {x: -31.982391, y: 2.5145874, z: 0.0000019073486} + m_Extent: {x: 51.12609, y: 46.604404, z: 27.749142} + m_DirtyAABB: 0 +--- !u!1 &6633355367928764342 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859542} + - component: {fileID: 6633355367931943586} + m_Layer: 0 + m_Name: Backpack1 + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859542 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764342} + m_LocalRotation: {x: 0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: -0, y: -138.49533, z: 34.48629} + m_LocalScale: {x: 0.01, y: 0.01, z: 0.01} + m_Children: [] + m_Father: {fileID: 6633355367928859588} + m_RootOrder: 2 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!137 &6633355367931943586 +SkinnedMeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764342} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 0 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 2100000, guid: 78660c09ff101ea4cb77de408a9527da, type: 2} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 0 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 + serializedVersion: 2 + m_Quality: 0 + m_UpdateWhenOffscreen: 0 + m_SkinnedMotionVectors: 1 + m_Mesh: {fileID: 4300004, guid: 860793eb3324391468f1c120a75ec049, type: 3} + m_Bones: + - {fileID: 6633355367928859552} + - {fileID: 6633355367928859622} + - {fileID: 6633355367928859538} + - {fileID: 6633355367928859590} + - {fileID: 6633355367928859544} + - {fileID: 6633355367928859556} + - {fileID: 6633355367928859554} + - {fileID: 6633355367928859606} + - {fileID: 6633355367928859646} + - {fileID: 6633355367928859568} + - {fileID: 6633355367928859548} + - {fileID: 6633355367928859594} + - {fileID: 6633355367928859598} + - {fileID: 6633355367928859586} + - {fileID: 6633355367928859574} + - {fileID: 6633355367928859562} + - {fileID: 6633355367928859566} + - {fileID: 6633355367928859628} + - {fileID: 6633355367928859616} + - {fileID: 6633355367928859620} + - {fileID: 6633355367928859610} + - {fileID: 6633355367928859614} + - {fileID: 6633355367928859602} + - {fileID: 6633355367928859604} + - {fileID: 6633355367928859644} + - {fileID: 6633355367928859582} + - {fileID: 6633355367928859546} + - {fileID: 6633355367928859592} + - {fileID: 6633355367928859596} + - {fileID: 6633355367928859584} + - {fileID: 6633355367928859572} + - {fileID: 6633355367928859560} + - {fileID: 6633355367928859564} + - {fileID: 6633355367928859626} + - {fileID: 6633355367928859630} + - {fileID: 6633355367928859618} + - {fileID: 6633355367928859608} + - {fileID: 6633355367928859612} + - {fileID: 6633355367928859600} + - {fileID: 6633355367928859528} + - {fileID: 6633355367928859524} + - {fileID: 6633355367928859520} + - {fileID: 6633355367928859522} + - {fileID: 6633355367928859642} + - {fileID: 6633355367928859580} + - {fileID: 6633355367928859536} + - {fileID: 6633355367928859624} + - {fileID: 6633355367928859636} + - {fileID: 6633355367928859640} + - {fileID: 6633355367928859578} + - {fileID: 6633355367928859550} + - {fileID: 6633355367928859638} + - {fileID: 6633355367928859634} + - {fileID: 6633355367928859532} + - {fileID: 6633355367928859576} + - {fileID: 6633355367928859632} + - {fileID: 6633355367928859534} + m_BlendShapeWeights: [] + m_RootBone: {fileID: 6633355367928859552} + m_AABB: + m_Center: {x: -28.89946, y: 44.769905, z: -1.5718498} + m_Extent: {x: 47.82257, y: 32.29384, z: 42.822285} + m_DirtyAABB: 0 +--- !u!1 &6633355367928764344 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859544} + m_Layer: 0 + m_Name: Head + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859544 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764344} + m_LocalRotation: {x: 0.071218304, y: -0.028954746, z: -0.059960153, w: 0.99523586} + m_LocalPosition: {x: -13.255574, y: -1.4210855e-14, z: 7.364747e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859556} + - {fileID: 6633355367928859554} + m_Father: {fileID: 6633355367928859590} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764346 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859546} + m_Layer: 0 + m_Name: Hand_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859546 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764346} + m_LocalRotation: {x: -0.0005926622, y: 0.085287675, z: -0.0069382307, w: 0.99633205} + m_LocalPosition: {x: -37.441414, y: 1.7053026e-13, z: -5.3290705e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859572} + - {fileID: 6633355367928859592} + - {fileID: 6633355367928859626} + m_Father: {fileID: 6633355367928859582} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764348 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859548} + m_Layer: 0 + m_Name: Hand_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859548 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764348} + m_LocalRotation: {x: 0.0005932963, y: -0.08528753, z: -0.006930911, w: 0.9963321} + m_LocalPosition: {x: -37.441498, y: 8.526513e-14, z: 2.7533531e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859574} + - {fileID: 6633355367928859594} + - {fileID: 6633355367928859628} + m_Father: {fileID: 6633355367928859568} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764350 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859550} + m_Layer: 0 + m_Name: Foot_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859550 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764350} + m_LocalRotation: {x: 0.81892335, y: 0.5656828, z: 0.024710972, w: -0.09357782} + m_LocalPosition: {x: -62.73464, y: -9.769963e-15, z: 2.4868996e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859638} + m_Father: {fileID: 6633355367928859578} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764352 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859552} + m_Layer: 0 + m_Name: Hips + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859552 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764352} + m_LocalRotation: {x: -0.4582423, y: 0.5385295, z: -0.4582423, w: 0.5385295} + m_LocalPosition: {x: -0, y: 1.0650933, z: 0.038611155} + m_LocalScale: {x: 0.01, y: 0.01, z: 0.01} + m_Children: + - {fileID: 6633355367928859534} + - {fileID: 6633355367928859532} + - {fileID: 6633355367928859622} + - {fileID: 6633355367928859642} + - {fileID: 6633355367928859640} + m_Father: {fileID: 6633355367928859588} + m_RootOrder: 5 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764354 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859554} + m_Layer: 0 + m_Name: Headgear_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859554 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764354} + m_LocalRotation: {x: -0.5234401, y: 0.5463846, z: 0.4754055, w: 0.44884726} + m_LocalPosition: {x: 1.3088212, y: 3.871166, z: 12.853171} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859544} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764356 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859556} + m_Layer: 0 + m_Name: Headgear_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859556 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764356} + m_LocalRotation: {x: 0.47540557, y: -0.44884732, z: 0.5234402, w: 0.54638463} + m_LocalPosition: {x: -8.526513e-14, y: -6.550316e-15, z: -13.487081} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859544} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764358 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859558} + - component: {fileID: 6633355367931943598} + m_Layer: 0 + m_Name: head1 + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859558 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764358} + m_LocalRotation: {x: 0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: -0, y: -176.9416, z: -6.1230974} + m_LocalScale: {x: 0.01, y: 0.01, z: 0.01} + m_Children: [] + m_Father: {fileID: 6633355367928859588} + m_RootOrder: 4 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!137 &6633355367931943598 +SkinnedMeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764358} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 0 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 2100000, guid: 78660c09ff101ea4cb77de408a9527da, type: 2} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 0 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 + serializedVersion: 2 + m_Quality: 0 + m_UpdateWhenOffscreen: 0 + m_SkinnedMotionVectors: 1 + m_Mesh: {fileID: 4300000, guid: 860793eb3324391468f1c120a75ec049, type: 3} + m_Bones: + - {fileID: 6633355367928859552} + - {fileID: 6633355367928859622} + - {fileID: 6633355367928859538} + - {fileID: 6633355367928859590} + - {fileID: 6633355367928859544} + - {fileID: 6633355367928859556} + - {fileID: 6633355367928859554} + - {fileID: 6633355367928859606} + - {fileID: 6633355367928859646} + - {fileID: 6633355367928859568} + - {fileID: 6633355367928859548} + - {fileID: 6633355367928859594} + - {fileID: 6633355367928859598} + - {fileID: 6633355367928859586} + - {fileID: 6633355367928859574} + - {fileID: 6633355367928859562} + - {fileID: 6633355367928859566} + - {fileID: 6633355367928859628} + - {fileID: 6633355367928859616} + - {fileID: 6633355367928859620} + - {fileID: 6633355367928859610} + - {fileID: 6633355367928859614} + - {fileID: 6633355367928859602} + - {fileID: 6633355367928859604} + - {fileID: 6633355367928859644} + - {fileID: 6633355367928859582} + - {fileID: 6633355367928859546} + - {fileID: 6633355367928859592} + - {fileID: 6633355367928859596} + - {fileID: 6633355367928859584} + - {fileID: 6633355367928859572} + - {fileID: 6633355367928859560} + - {fileID: 6633355367928859564} + - {fileID: 6633355367928859626} + - {fileID: 6633355367928859630} + - {fileID: 6633355367928859618} + - {fileID: 6633355367928859608} + - {fileID: 6633355367928859612} + - {fileID: 6633355367928859600} + - {fileID: 6633355367928859528} + - {fileID: 6633355367928859524} + - {fileID: 6633355367928859520} + - {fileID: 6633355367928859522} + - {fileID: 6633355367928859642} + - {fileID: 6633355367928859580} + - {fileID: 6633355367928859536} + - {fileID: 6633355367928859624} + - {fileID: 6633355367928859636} + - {fileID: 6633355367928859640} + - {fileID: 6633355367928859578} + - {fileID: 6633355367928859550} + - {fileID: 6633355367928859638} + - {fileID: 6633355367928859634} + - {fileID: 6633355367928859532} + - {fileID: 6633355367928859576} + - {fileID: 6633355367928859632} + - {fileID: 6633355367928859534} + m_BlendShapeWeights: [] + m_RootBone: {fileID: 6633355367928859552} + m_AABB: + m_Center: {x: -71.52329, y: 10.616158, z: -1.0935726} + m_Extent: {x: 26.513672, y: 26.529358, z: 17.550735} + m_DirtyAABB: 0 +--- !u!1 &6633355367928764360 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859560} + m_Layer: 0 + m_Name: Index_Intermediate_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859560 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764360} + m_LocalRotation: {x: -0.017963478, y: -0.041126773, z: 0.25885317, w: 0.96487355} + m_LocalPosition: {x: -10.023805, y: -1.4210855e-13, z: -1.7763568e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859564} + m_Father: {fileID: 6633355367928859572} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764362 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859562} + m_Layer: 0 + m_Name: Index_Intermediate_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859562 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764362} + m_LocalRotation: {x: 0.017964862, y: 0.041127786, z: 0.25889415, w: 0.96486247} + m_LocalPosition: {x: -10.023557, y: -1.4210855e-13, z: 8.881784e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859566} + m_Father: {fileID: 6633355367928859574} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764364 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859564} + m_Layer: 0 + m_Name: Index_Distal_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859564 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764364} + m_LocalRotation: {x: 0.012835801, y: 0.013473716, z: 0.08711711, w: 0.99602425} + m_LocalPosition: {x: -7.9785438, y: 1.7053026e-13, z: 4.440892e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859560} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764366 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859566} + m_Layer: 0 + m_Name: Index_Distal_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859566 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764366} + m_LocalRotation: {x: -0.013842603, y: -0.01243883, z: 0.16401216, w: 0.98628277} + m_LocalPosition: {x: -7.9783425, y: 5.684342e-14, z: -1.7763568e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859562} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764368 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859568} + m_Layer: 0 + m_Name: LowerArm_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859568 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764368} + m_LocalRotation: {x: -1.11818e-29, y: 0.049585894, z: 1.0904029e-28, w: 0.9987699} + m_LocalPosition: {x: -33.66694, y: 2.842171e-14, z: 4.2188475e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859548} + m_Father: {fileID: 6633355367928859646} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764370 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859570} + - component: {fileID: 6633355367931943596} + m_Layer: 0 + m_Name: Leg1 + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859570 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764370} + m_LocalRotation: {x: 0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: -0, y: -96.69518, z: -8.197696} + m_LocalScale: {x: 0.01, y: 0.01, z: 0.01} + m_Children: [] + m_Father: {fileID: 6633355367928859588} + m_RootOrder: 6 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!137 &6633355367931943596 +SkinnedMeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764370} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 0 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 2100000, guid: 78660c09ff101ea4cb77de408a9527da, type: 2} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 0 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 + serializedVersion: 2 + m_Quality: 0 + m_UpdateWhenOffscreen: 0 + m_SkinnedMotionVectors: 1 + m_Mesh: {fileID: 4300008, guid: 860793eb3324391468f1c120a75ec049, type: 3} + m_Bones: + - {fileID: 6633355367928859552} + - {fileID: 6633355367928859622} + - {fileID: 6633355367928859538} + - {fileID: 6633355367928859590} + - {fileID: 6633355367928859544} + - {fileID: 6633355367928859556} + - {fileID: 6633355367928859554} + - {fileID: 6633355367928859606} + - {fileID: 6633355367928859646} + - {fileID: 6633355367928859568} + - {fileID: 6633355367928859548} + - {fileID: 6633355367928859594} + - {fileID: 6633355367928859598} + - {fileID: 6633355367928859586} + - {fileID: 6633355367928859574} + - {fileID: 6633355367928859562} + - {fileID: 6633355367928859566} + - {fileID: 6633355367928859628} + - {fileID: 6633355367928859616} + - {fileID: 6633355367928859620} + - {fileID: 6633355367928859610} + - {fileID: 6633355367928859614} + - {fileID: 6633355367928859602} + - {fileID: 6633355367928859604} + - {fileID: 6633355367928859644} + - {fileID: 6633355367928859582} + - {fileID: 6633355367928859546} + - {fileID: 6633355367928859592} + - {fileID: 6633355367928859596} + - {fileID: 6633355367928859584} + - {fileID: 6633355367928859572} + - {fileID: 6633355367928859560} + - {fileID: 6633355367928859564} + - {fileID: 6633355367928859626} + - {fileID: 6633355367928859630} + - {fileID: 6633355367928859618} + - {fileID: 6633355367928859608} + - {fileID: 6633355367928859612} + - {fileID: 6633355367928859600} + - {fileID: 6633355367928859528} + - {fileID: 6633355367928859524} + - {fileID: 6633355367928859520} + - {fileID: 6633355367928859522} + - {fileID: 6633355367928859642} + - {fileID: 6633355367928859580} + - {fileID: 6633355367928859536} + - {fileID: 6633355367928859624} + - {fileID: 6633355367928859636} + - {fileID: 6633355367928859640} + - {fileID: 6633355367928859578} + - {fileID: 6633355367928859550} + - {fileID: 6633355367928859638} + - {fileID: 6633355367928859634} + - {fileID: 6633355367928859532} + - {fileID: 6633355367928859576} + - {fileID: 6633355367928859632} + - {fileID: 6633355367928859534} + m_BlendShapeWeights: [] + m_RootBone: {fileID: 6633355367928859552} + m_AABB: + m_Center: {x: 53.24073, y: -15.325765, z: 0.17397118} + m_Extent: {x: 69.29029, y: 36.12146, z: 51.77101} + m_DirtyAABB: 0 +--- !u!1 &6633355367928764372 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859572} + m_Layer: 0 + m_Name: Index_Proximal_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859572 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764372} + m_LocalRotation: {x: -0.0028644414, y: -0.042128764, z: 0.08165133, w: 0.99576604} + m_LocalPosition: {x: -13.357139, y: -0.09103328, z: -13.440055} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859560} + m_Father: {fileID: 6633355367928859546} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764374 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859574} + m_Layer: 0 + m_Name: Index_Proximal_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859574 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764374} + m_LocalRotation: {x: 0.0028642316, y: 0.04212831, z: 0.08163239, w: 0.9957676} + m_LocalPosition: {x: -13.357445, y: -0.09144714, z: 13.440094} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859562} + m_Father: {fileID: 6633355367928859548} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764376 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859576} + m_Layer: 0 + m_Name: magazine_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859576 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764376} + m_LocalRotation: {x: -0.4881349, y: 0.5173824, z: 0.48743477, w: 0.5064061} + m_LocalPosition: {x: -0.004472253, y: -40.147858, z: 0.03500123} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859532} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764378 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859578} + m_Layer: 0 + m_Name: LowerLeg_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859578 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764378} + m_LocalRotation: {x: 0.66088396, y: 0.05163587, z: -0.04218814, w: 0.74752015} + m_LocalPosition: {x: -33.437313, y: 4.973799e-14, z: -8.881784e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859550} + m_Father: {fileID: 6633355367928859640} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764380 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859580} + m_Layer: 0 + m_Name: LowerLeg_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859580 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764380} + m_LocalRotation: {x: -0.6640286, y: -0.051458772, z: -0.042404156, w: 0.74472815} + m_LocalPosition: {x: -33.436848, y: -2.4868996e-14, z: 6.217249e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859536} + m_Father: {fileID: 6633355367928859642} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764382 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859582} + m_Layer: 0 + m_Name: LowerArm_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859582 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764382} + m_LocalRotation: {x: 4.9720758e-29, y: -0.049585804, z: 1.0024346e-27, w: 0.9987699} + m_LocalPosition: {x: -33.666927, y: 0, z: -1.4432899e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859546} + m_Father: {fileID: 6633355367928859644} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764384 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859584} + m_Layer: 0 + m_Name: RestOfFingers_Distal_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859584 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764384} + m_LocalRotation: {x: 0.0041774614, y: 0.0038508072, z: 0.015434083, w: 0.99986476} + m_LocalPosition: {x: -8.174185, y: 2.842171e-14, z: 2.6645353e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859596} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764386 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859586} + m_Layer: 0 + m_Name: RestOfFingers_Distal_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859586 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764386} + m_LocalRotation: {x: -0.004532425, y: -0.003425033, z: 0.11285209, w: 0.9935956} + m_LocalPosition: {x: -8.173605, y: -8.526513e-14, z: -3.5527137e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859598} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764388 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859588} + m_Layer: 0 + m_Name: Character + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859588 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764388} + m_LocalRotation: {x: 0, y: 0, z: 0, w: 1} + m_LocalPosition: {x: 44.336212, y: 0.6996672, z: -10.997339} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859526} + - {fileID: 6633355367928859530} + - {fileID: 6633355367928859542} + - {fileID: 6633355367928859540} + - {fileID: 6633355367928859558} + - {fileID: 6633355367928859552} + - {fileID: 6633355367928859570} + m_Father: {fileID: 0} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764390 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859590} + m_Layer: 0 + m_Name: Neck + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859590 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764390} + m_LocalRotation: {x: 0.9631089, y: -0.26911193, z: 1.6478353e-17, w: -5.897341e-17} + m_LocalPosition: {x: -17.7523, y: -2.842171e-14, z: 0.000000026610966} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859544} + m_Father: {fileID: 6633355367928859538} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764392 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859592} + m_Layer: 0 + m_Name: RestOfFingers_Proximal_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859592 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764392} + m_LocalRotation: {x: -0.0028322286, y: -0.060177784, z: 0.060808677, w: 0.9963297} + m_LocalPosition: {x: -18.094389, y: 8.526513e-14, z: -1.0658141e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859596} + m_Father: {fileID: 6633355367928859546} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764394 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859594} + m_Layer: 0 + m_Name: RestOfFingers_Proximal_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859594 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764394} + m_LocalRotation: {x: 0.0028340812, y: 0.060178064, z: 0.060824323, w: 0.9963288} + m_LocalPosition: {x: -18.094444, y: 5.684342e-14, z: 1.7763568e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859598} + m_Father: {fileID: 6633355367928859548} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764396 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859596} + m_Layer: 0 + m_Name: RestOfFingers_Intermediate_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859596 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764396} + m_LocalRotation: {x: -0.0061721927, y: -0.012479491, z: 0.3441051, w: 0.93882793} + m_LocalPosition: {x: -8.025833, y: -1.7053026e-13, z: 4.8849813e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859584} + m_Father: {fileID: 6633355367928859592} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764398 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859598} + m_Layer: 0 + m_Name: RestOfFingers_Intermediate_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859598 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764398} + m_LocalRotation: {x: 0.0061717043, y: 0.012478555, z: 0.3440602, w: 0.9388444} + m_LocalPosition: {x: -8.02586, y: -1.1368684e-13, z: -5.3290705e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859586} + m_Father: {fileID: 6633355367928859594} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764400 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859600} + m_Layer: 0 + m_Name: ShoulderPadBlade_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859600 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764400} + m_LocalRotation: {x: -0.0066352007, y: 0.011333432, z: 0.50519216, w: 0.86290693} + m_LocalPosition: {x: -13.296815, y: -22.177006, z: 0.5073527} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859608} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764402 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859602} + m_Layer: 0 + m_Name: ShoulderPadBlade_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859602 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764402} + m_LocalRotation: {x: 0.0066350633, y: -0.011333145, z: 0.5051939, w: 0.862906} + m_LocalPosition: {x: -13.296959, y: -22.176832, z: -0.5073402} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859610} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764404 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859604} + m_Layer: 0 + m_Name: Shoulder_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859604 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764404} + m_LocalRotation: {x: -0.29884863, y: 0.73412436, z: -0.41826612, w: -0.4436265} + m_LocalPosition: {x: -12.348376, y: -6.29761, z: -18.7061} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859608} + - {fileID: 6633355367928859644} + m_Father: {fileID: 6633355367928859538} + m_RootOrder: 3 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764406 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859606} + m_Layer: 0 + m_Name: Shoulder_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859606 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764406} + m_LocalRotation: {x: -0.2988425, y: 0.7341269, z: 0.41826975, w: 0.4436229} + m_LocalPosition: {x: -12.348807, y: -6.29741, z: 18.706108} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859610} + - {fileID: 6633355367928859646} + m_Father: {fileID: 6633355367928859538} + m_RootOrder: 2 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764408 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859608} + m_Layer: 0 + m_Name: ShoulderPadCTRL_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859608 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764408} + m_LocalRotation: {x: -0.039328646, y: 0.10945006, z: -0.6506794, w: 0.7503934} + m_LocalPosition: {x: -9.938715, y: 14.657999, z: 1.039447} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859600} + - {fileID: 6633355367928859612} + m_Father: {fileID: 6633355367928859604} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764410 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859610} + m_Layer: 0 + m_Name: ShoulderPadCTRL_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859610 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764410} + m_LocalRotation: {x: 0.039328095, y: -0.109450735, z: -0.65068716, w: 0.75038666} + m_LocalPosition: {x: -9.938806, y: 14.657373, z: -1.0394562} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 6633355367928859602} + - {fileID: 6633355367928859614} + m_Father: {fileID: 6633355367928859606} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764412 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859612} + m_Layer: 0 + m_Name: ShoulderPadBody_Right + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859612 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764412} + m_LocalRotation: {x: -0.0066352007, y: 0.011333432, z: 0.50519216, w: 0.86290693} + m_LocalPosition: {x: -16.02375, y: -6.3948846e-14, z: 1.2878587e-14} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859608} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &6633355367928764414 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 6633355367928859614} + m_Layer: 0 + m_Name: ShoulderPadBody_Left + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &6633355367928859614 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 6633355367928764414} + m_LocalRotation: {x: 0.0066350633, y: -0.011333145, z: 0.5051939, w: 0.862906} + m_LocalPosition: {x: -16.023825, y: -1.4210855e-14, z: 2.6645353e-15} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 6633355367928859610} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} diff --git a/Assets/ML-Agents/Timers/dont touch me plz_timers.json.meta b/Assets/Prefabs/CharacterPrefabs/Character.prefab.meta old mode 100755 new mode 100644 similarity index 62% rename from Assets/ML-Agents/Timers/dont touch me plz_timers.json.meta rename to Assets/Prefabs/CharacterPrefabs/Character.prefab.meta index db7171f..4a4509a --- a/Assets/ML-Agents/Timers/dont touch me plz_timers.json.meta +++ b/Assets/Prefabs/CharacterPrefabs/Character.prefab.meta @@ -1,6 +1,6 @@ fileFormatVersion: 2 -guid: 249f8a9af2b0bc041a08a0009a6fdf44 -TextScriptImporter: +guid: 0d0b10717aa98fe44bcf651ad0901a89 +PrefabImporter: externalObjects: {} userData: assetBundleName: diff --git a/Assets/Prefabs/Player.prefab b/Assets/Prefabs/CharacterPrefabs/Player.prefab old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/Player.prefab rename to Assets/Prefabs/CharacterPrefabs/Player.prefab diff --git a/Assets/Prefabs/Player.prefab.meta b/Assets/Prefabs/CharacterPrefabs/Player.prefab.meta similarity index 100% rename from Assets/Prefabs/Player.prefab.meta rename to Assets/Prefabs/CharacterPrefabs/Player.prefab.meta diff --git a/Assets/Prefabs/EntityPrefabs.meta b/Assets/Prefabs/EntityPrefabs.meta new file mode 100644 index 0000000..a069136 --- /dev/null +++ b/Assets/Prefabs/EntityPrefabs.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: 61923742bbdb59f45816f6d760980fc4 +folderAsset: yes +DefaultImporter: + externalObjects: {} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Prefabs/Flag zone.prefab b/Assets/Prefabs/EntityPrefabs/Flag zone.prefab old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/Flag zone.prefab rename to Assets/Prefabs/EntityPrefabs/Flag zone.prefab diff --git a/Assets/Prefabs/Flag zone.prefab.meta b/Assets/Prefabs/EntityPrefabs/Flag zone.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/Flag zone.prefab.meta rename to Assets/Prefabs/EntityPrefabs/Flag zone.prefab.meta diff --git a/Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab b/Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab new file mode 100644 index 0000000..afd2291 --- /dev/null +++ b/Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab @@ -0,0 +1,49 @@ +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!1 &7378066357312323927 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 7378066357312323926} + - component: {fileID: 7378066357312323925} + m_Layer: 0 + m_Name: navPointPrefab + m_TagString: Point + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &7378066357312323926 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7378066357312323927} + m_LocalRotation: {x: -0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: 0, y: 0, z: 0} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 0} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!114 &7378066357312323925 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7378066357312323927} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} + m_Name: + m_EditorClassIdentifier: + navType: 1 + PointId: 0 + DeathAttr: 0 + EnemiesSeen: [] diff --git a/Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab.meta b/Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab.meta new file mode 100644 index 0000000..f826d76 --- /dev/null +++ b/Assets/Prefabs/EntityPrefabs/navPointPrefab.prefab.meta @@ -0,0 +1,7 @@ +fileFormatVersion: 2 +guid: f58a14c8ba97dfa4f9f4103bbee852d4 +PrefabImporter: + externalObjects: {} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Prefabs/PickupPrefabs.meta b/Assets/Prefabs/PickupPrefabs.meta new file mode 100644 index 0000000..b406ea9 --- /dev/null +++ b/Assets/Prefabs/PickupPrefabs.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: 132a5e983df3eaa41bbb1e9bda3dd97e +folderAsset: yes +DefaultImporter: + externalObjects: {} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Prefabs/ammo.prefab b/Assets/Prefabs/PickupPrefabs/ammo.prefab old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/ammo.prefab rename to Assets/Prefabs/PickupPrefabs/ammo.prefab diff --git a/Assets/Prefabs/ammo.prefab.meta b/Assets/Prefabs/PickupPrefabs/ammo.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/ammo.prefab.meta rename to Assets/Prefabs/PickupPrefabs/ammo.prefab.meta diff --git a/Assets/Prefabs/armour.prefab b/Assets/Prefabs/PickupPrefabs/armour.prefab old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/armour.prefab rename to Assets/Prefabs/PickupPrefabs/armour.prefab diff --git a/Assets/Prefabs/armour.prefab.meta b/Assets/Prefabs/PickupPrefabs/armour.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/armour.prefab.meta rename to Assets/Prefabs/PickupPrefabs/armour.prefab.meta diff --git a/Assets/Prefabs/health.prefab b/Assets/Prefabs/PickupPrefabs/health.prefab old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/health.prefab rename to Assets/Prefabs/PickupPrefabs/health.prefab diff --git a/Assets/Prefabs/health.prefab.meta b/Assets/Prefabs/PickupPrefabs/health.prefab.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Prefabs/health.prefab.meta rename to Assets/Prefabs/PickupPrefabs/health.prefab.meta diff --git a/Assets/Resources.meta b/Assets/Resources.meta new file mode 100644 index 0000000..f303094 --- /dev/null +++ b/Assets/Resources.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: 61178e06e35bbb84daabda8d15a50d45 +folderAsset: yes +DefaultImporter: + externalObjects: {} + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Resources/BillingMode.json b/Assets/Resources/BillingMode.json new file mode 100644 index 0000000..6f4bfb7 --- /dev/null +++ b/Assets/Resources/BillingMode.json @@ -0,0 +1 @@ +{"androidStore":"GooglePlay"} \ No newline at end of file diff --git a/Assets/ML-Agents/Timers/DemoScene01_timers.json.meta b/Assets/Resources/BillingMode.json.meta old mode 100755 new mode 100644 similarity index 75% rename from Assets/ML-Agents/Timers/DemoScene01_timers.json.meta rename to Assets/Resources/BillingMode.json.meta index fe18e4a..255ee91 --- a/Assets/ML-Agents/Timers/DemoScene01_timers.json.meta +++ b/Assets/Resources/BillingMode.json.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: 1b328c4e26e7a994e8a42f269ca7419b +guid: 01004bca71fbe914d9e15e843ac4aca0 TextScriptImporter: externalObjects: {} userData: diff --git a/Assets/Scenes/AnimationTestScene.unity b/Assets/Scenes/AnimationTestScene.unity new file mode 100644 index 0000000..b3ec676 --- /dev/null +++ b/Assets/Scenes/AnimationTestScene.unity @@ -0,0 +1,1060 @@ +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!29 &1 +OcclusionCullingSettings: + m_ObjectHideFlags: 0 + serializedVersion: 2 + m_OcclusionBakeSettings: + smallestOccluder: 5 + smallestHole: 0.25 + backfaceThreshold: 100 + m_SceneGUID: 00000000000000000000000000000000 + m_OcclusionCullingData: {fileID: 0} +--- !u!104 &2 +RenderSettings: + m_ObjectHideFlags: 0 + serializedVersion: 9 + m_Fog: 0 + m_FogColor: {r: 0.5, g: 0.5, b: 0.5, a: 1} + m_FogMode: 3 + m_FogDensity: 0.01 + m_LinearFogStart: 0 + m_LinearFogEnd: 300 + m_AmbientSkyColor: {r: 0.212, g: 0.227, b: 0.259, a: 1} + m_AmbientEquatorColor: {r: 0.114, g: 0.125, b: 0.133, a: 1} + m_AmbientGroundColor: {r: 0.047, g: 0.043, b: 0.035, a: 1} + m_AmbientIntensity: 1 + m_AmbientMode: 0 + m_SubtractiveShadowColor: {r: 0.42, g: 0.478, b: 0.627, a: 1} + m_SkyboxMaterial: {fileID: 10304, guid: 0000000000000000f000000000000000, type: 0} + m_HaloStrength: 0.5 + m_FlareStrength: 1 + m_FlareFadeSpeed: 3 + m_HaloTexture: {fileID: 0} + m_SpotCookie: {fileID: 10001, guid: 0000000000000000e000000000000000, type: 0} + m_DefaultReflectionMode: 0 + m_DefaultReflectionResolution: 128 + m_ReflectionBounces: 1 + m_ReflectionIntensity: 1 + m_CustomReflection: {fileID: 0} + m_Sun: {fileID: 0} + m_IndirectSpecularColor: {r: 0.44657898, g: 0.4964133, b: 0.5748178, a: 1} + m_UseRadianceAmbientProbe: 0 +--- !u!157 &3 +LightmapSettings: + m_ObjectHideFlags: 0 + serializedVersion: 11 + m_GIWorkflowMode: 1 + m_GISettings: + serializedVersion: 2 + m_BounceScale: 1 + m_IndirectOutputScale: 1 + m_AlbedoBoost: 1 + m_EnvironmentLightingMode: 0 + m_EnableBakedLightmaps: 1 + m_EnableRealtimeLightmaps: 0 + m_LightmapEditorSettings: + serializedVersion: 12 + m_Resolution: 2 + m_BakeResolution: 40 + m_AtlasSize: 1024 + m_AO: 0 + m_AOMaxDistance: 1 + m_CompAOExponent: 1 + m_CompAOExponentDirect: 0 + m_ExtractAmbientOcclusion: 0 + m_Padding: 2 + m_LightmapParameters: {fileID: 0} + m_LightmapsBakeMode: 1 + m_TextureCompression: 1 + m_FinalGather: 0 + m_FinalGatherFiltering: 1 + m_FinalGatherRayCount: 256 + m_ReflectionCompression: 2 + m_MixedBakeMode: 2 + m_BakeBackend: 1 + m_PVRSampling: 1 + m_PVRDirectSampleCount: 32 + m_PVRSampleCount: 512 + m_PVRBounces: 2 + m_PVREnvironmentSampleCount: 256 + m_PVREnvironmentReferencePointCount: 2048 + m_PVRFilteringMode: 1 + m_PVRDenoiserTypeDirect: 1 + m_PVRDenoiserTypeIndirect: 1 + m_PVRDenoiserTypeAO: 1 + m_PVRFilterTypeDirect: 0 + m_PVRFilterTypeIndirect: 0 + m_PVRFilterTypeAO: 0 + m_PVREnvironmentMIS: 1 + m_PVRCulling: 1 + m_PVRFilteringGaussRadiusDirect: 1 + m_PVRFilteringGaussRadiusIndirect: 5 + m_PVRFilteringGaussRadiusAO: 2 + m_PVRFilteringAtrousPositionSigmaDirect: 0.5 + m_PVRFilteringAtrousPositionSigmaIndirect: 2 + m_PVRFilteringAtrousPositionSigmaAO: 1 + m_ExportTrainingData: 0 + m_TrainingDataDestination: TrainingData + m_LightProbeSampleCountMultiplier: 4 + m_LightingDataAsset: {fileID: 0} + m_UseShadowmask: 1 +--- !u!196 &4 +NavMeshSettings: + serializedVersion: 2 + m_ObjectHideFlags: 0 + m_BuildSettings: + serializedVersion: 2 + agentTypeID: 0 + agentRadius: 0.5 + agentHeight: 2 + agentSlope: 45 + agentClimb: 0.4 + ledgeDropHeight: 0 + maxJumpAcrossDistance: 0 + minRegionArea: 2 + manualCellSize: 0 + cellSize: 0.16666667 + manualTileSize: 0 + tileSize: 256 + accuratePlacement: 0 + debug: + m_Flags: 0 + m_NavMeshData: {fileID: 0} +--- !u!1 &93904468 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 93904472} + - component: {fileID: 93904471} + - component: {fileID: 93904470} + - component: {fileID: 93904469} + m_Layer: 0 + m_Name: Plane (1) + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!64 &93904469 +MeshCollider: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 93904468} + m_Material: {fileID: 0} + m_IsTrigger: 0 + m_Enabled: 1 + serializedVersion: 4 + m_Convex: 0 + m_CookingOptions: 30 + m_Mesh: {fileID: 10209, guid: 0000000000000000e000000000000000, type: 0} +--- !u!23 &93904470 +MeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 93904468} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 2 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 10303, guid: 0000000000000000f000000000000000, type: 0} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 1 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 +--- !u!33 &93904471 +MeshFilter: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 93904468} + m_Mesh: {fileID: 10209, guid: 0000000000000000e000000000000000, type: 0} +--- !u!4 &93904472 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 93904468} + m_LocalRotation: {x: 0, y: 0, z: 0, w: 1} + m_LocalPosition: {x: 0, y: -0.53, z: 10.04} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 0} + m_RootOrder: 4 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &738827739 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 738827741} + - component: {fileID: 738827740} + m_Layer: 0 + m_Name: Camera Holder + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!114 &738827740 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 738827739} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 3d606407023147d7b4d530a9593e9697, type: 3} + m_Name: + m_EditorClassIdentifier: + targetTransform: {fileID: 7072444062111317995} + cameraTransform: {fileID: 1403458766} + cameraPivotTransform: {fileID: 1054021701} + minimumPivot: -35 + maximumPivot: 35 +--- !u!4 &738827741 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 738827739} + m_LocalRotation: {x: 0, y: 0, z: 0, w: 1} + m_LocalPosition: {x: 0, y: 0, z: 0} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 1054021701} + m_Father: {fileID: 0} + m_RootOrder: 3 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &1054021700 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 1054021701} + m_Layer: 0 + m_Name: Camera Pivot + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!4 &1054021701 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1054021700} + m_LocalRotation: {x: -0, y: -0, z: -0, w: 1} + m_LocalPosition: {x: 0, y: 0, z: 0} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: + - {fileID: 1403458766} + m_Father: {fileID: 738827741} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &1225778554 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 1225778558} + - component: {fileID: 1225778557} + - component: {fileID: 1225778556} + - component: {fileID: 1225778555} + m_Layer: 0 + m_Name: Plane + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!64 &1225778555 +MeshCollider: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1225778554} + m_Material: {fileID: 0} + m_IsTrigger: 0 + m_Enabled: 1 + serializedVersion: 4 + m_Convex: 0 + m_CookingOptions: 30 + m_Mesh: {fileID: 10209, guid: 0000000000000000e000000000000000, type: 0} +--- !u!23 &1225778556 +MeshRenderer: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1225778554} + m_Enabled: 1 + m_CastShadows: 1 + m_ReceiveShadows: 1 + m_DynamicOccludee: 1 + m_MotionVectors: 1 + m_LightProbeUsage: 1 + m_ReflectionProbeUsage: 1 + m_RayTracingMode: 2 + m_RenderingLayerMask: 1 + m_RendererPriority: 0 + m_Materials: + - {fileID: 10303, guid: 0000000000000000f000000000000000, type: 0} + m_StaticBatchInfo: + firstSubMesh: 0 + subMeshCount: 0 + m_StaticBatchRoot: {fileID: 0} + m_ProbeAnchor: {fileID: 0} + m_LightProbeVolumeOverride: {fileID: 0} + m_ScaleInLightmap: 1 + m_ReceiveGI: 1 + m_PreserveUVs: 0 + m_IgnoreNormalsForChartDetection: 0 + m_ImportantGI: 0 + m_StitchLightmapSeams: 1 + m_SelectedEditorRenderState: 3 + m_MinimumChartSize: 4 + m_AutoUVMaxDistance: 0.5 + m_AutoUVMaxAngle: 89 + m_LightmapParameters: {fileID: 0} + m_SortingLayerID: 0 + m_SortingLayer: 0 + m_SortingOrder: 0 +--- !u!33 &1225778557 +MeshFilter: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1225778554} + m_Mesh: {fileID: 10209, guid: 0000000000000000e000000000000000, type: 0} +--- !u!4 &1225778558 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1225778554} + m_LocalRotation: {x: 0, y: 0, z: 0, w: 1} + m_LocalPosition: {x: 0, y: 0, z: 0} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 0} + m_RootOrder: 2 + m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} +--- !u!1 &1403458763 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 1403458766} + - component: {fileID: 1403458764} + m_Layer: 0 + m_Name: Main Camera + m_TagString: MainCamera + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!20 &1403458764 +Camera: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1403458763} + m_Enabled: 1 + serializedVersion: 2 + m_ClearFlags: 1 + m_BackGroundColor: {r: 0.19215687, g: 0.3019608, b: 0.4745098, a: 0} + m_projectionMatrixMode: 1 + m_GateFitMode: 2 + m_FOVAxisMode: 0 + m_SensorSize: {x: 36, y: 24} + m_LensShift: {x: 0, y: 0} + m_FocalLength: 50 + m_NormalizedViewPortRect: + serializedVersion: 2 + x: 0 + y: 0 + width: 1 + height: 1 + near clip plane: 0.3 + far clip plane: 1000 + field of view: 26.991467 + orthographic: 0 + orthographic size: 5 + m_Depth: 0 + m_CullingMask: + serializedVersion: 2 + m_Bits: 4294967295 + m_RenderingPath: -1 + m_TargetTexture: {fileID: 0} + m_TargetDisplay: 0 + m_TargetEye: 3 + m_HDR: 1 + m_AllowMSAA: 1 + m_AllowDynamicResolution: 0 + m_ForceIntoRT: 0 + m_OcclusionCulling: 1 + m_StereoConvergence: 10 + m_StereoSeparation: 0.022 +--- !u!4 &1403458766 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1403458763} + m_LocalRotation: {x: 0.08193846, y: 0, z: 0, w: 0.9966374} + m_LocalPosition: {x: 0, y: 3.39, z: -10.29} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 1054021701} + m_RootOrder: 0 + m_LocalEulerAnglesHint: {x: 9.4, y: 0, z: 0} +--- !u!1 &1755358664 +GameObject: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + serializedVersion: 6 + m_Component: + - component: {fileID: 1755358666} + - component: {fileID: 1755358665} + m_Layer: 0 + m_Name: Directional Light + m_TagString: Untagged + m_Icon: {fileID: 0} + m_NavMeshLayer: 0 + m_StaticEditorFlags: 0 + m_IsActive: 1 +--- !u!108 &1755358665 +Light: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1755358664} + m_Enabled: 1 + serializedVersion: 10 + m_Type: 1 + m_Shape: 0 + m_Color: {r: 1, g: 0.95686275, b: 0.8392157, a: 1} + m_Intensity: 1 + m_Range: 10 + m_SpotAngle: 30 + m_InnerSpotAngle: 21.80208 + m_CookieSize: 10 + m_Shadows: + m_Type: 2 + m_Resolution: -1 + m_CustomResolution: -1 + m_Strength: 1 + m_Bias: 0.05 + m_NormalBias: 0.4 + m_NearPlane: 0.2 + m_CullingMatrixOverride: + e00: 1 + e01: 0 + e02: 0 + e03: 0 + e10: 0 + e11: 1 + e12: 0 + e13: 0 + e20: 0 + e21: 0 + e22: 1 + e23: 0 + e30: 0 + e31: 0 + e32: 0 + e33: 1 + m_UseCullingMatrixOverride: 0 + m_Cookie: {fileID: 0} + m_DrawHalo: 0 + m_Flare: {fileID: 0} + m_RenderMode: 0 + m_CullingMask: + serializedVersion: 2 + m_Bits: 4294967295 + m_RenderingLayerMask: 1 + m_Lightmapping: 4 + m_LightShadowCasterMode: 0 + m_AreaSize: {x: 1, y: 1} + m_BounceIntensity: 1 + m_ColorTemperature: 6570 + m_UseColorTemperature: 0 + m_BoundingSphereOverride: {x: 0, y: 0, z: 0, w: 0} + m_UseBoundingSphereOverride: 0 + m_ShadowRadius: 0 + m_ShadowAngle: 0 +--- !u!4 &1755358666 +Transform: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 1755358664} + m_LocalRotation: {x: 0.40821788, y: -0.23456968, z: 0.10938163, w: 0.8754261} + m_LocalPosition: {x: 0, y: 3, z: 0} + m_LocalScale: {x: 1, y: 1, z: 1} + m_Children: [] + m_Father: {fileID: 0} + m_RootOrder: 1 + m_LocalEulerAnglesHint: {x: 50, y: -30, z: 0} +--- !u!1001 &4478891302374877197 +PrefabInstance: + m_ObjectHideFlags: 0 + serializedVersion: 2 + m_Modification: + m_TransformParent: {fileID: 0} + m_Modifications: + - target: {fileID: 6633355367928763904, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763906, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763908, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763910, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763912, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763914, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763916, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763918, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763920, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763922, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763924, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763926, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763928, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763930, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763932, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928763934, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764320, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764322, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764324, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764326, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764328, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764330, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764332, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764334, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764336, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764338, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764340, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764342, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764344, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764346, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764348, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764350, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764352, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764354, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764356, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764358, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764360, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764362, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764364, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764366, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764368, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764370, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764372, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764374, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764376, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764378, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764380, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764382, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764384, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764386, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764388, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Name + value: Character + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764388, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764390, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764392, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764394, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764396, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764398, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764400, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764402, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764404, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764406, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764408, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764410, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764412, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928764414, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_Layer + value: 10 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_RootOrder + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalPosition.x + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalPosition.y + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalPosition.z + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalRotation.w + value: 1 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalRotation.x + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalRotation.y + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalRotation.z + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalEulerAnglesHint.x + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalEulerAnglesHint.y + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + propertyPath: m_LocalEulerAnglesHint.z + value: 0 + objectReference: {fileID: 0} + m_RemovedComponents: [] + m_SourcePrefab: {fileID: 100100000, guid: 0d0b10717aa98fe44bcf651ad0901a89, type: 3} +--- !u!1 &7072444062111317993 stripped +GameObject: + m_CorrespondingSourceObject: {fileID: 6633355367928764388, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + m_PrefabInstance: {fileID: 4478891302374877197} + m_PrefabAsset: {fileID: 0} +--- !u!114 &7072444062111317994 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444062111317993} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: aeafb7b8074141969e8779cd3d4a9d08, type: 3} + m_Name: + m_EditorClassIdentifier: + myTransform: {fileID: 0} + myAnimatorHandler: {fileID: 0} + myRigidbody: {fileID: 0} + jumpForce: 1800 + normalCamera: {fileID: 0} + movementSpeed: 5 + rotationSpeed: 10 +--- !u!4 &7072444062111317995 stripped +Transform: + m_CorrespondingSourceObject: {fileID: 6633355367928859588, guid: 0d0b10717aa98fe44bcf651ad0901a89, + type: 3} + m_PrefabInstance: {fileID: 4478891302374877197} + m_PrefabAsset: {fileID: 0} +--- !u!114 &7072444062111317996 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444062111317993} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: 77c38ddfaba349c590d4a6583f7efac4, type: 3} + m_Name: + m_EditorClassIdentifier: + horizontal: 0 + vertical: 0 + moveAmount: 0 + mouseX: 0 + mouseY: 0 + crouchPressed: 0 + jumpPressed: 0 + firePressed: 0 +--- !u!54 &7072444062111317998 +Rigidbody: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444062111317993} + serializedVersion: 2 + m_Mass: 1 + m_Drag: 8 + m_AngularDrag: 20 + m_UseGravity: 1 + m_IsKinematic: 0 + m_Interpolate: 0 + m_Constraints: 80 + m_CollisionDetection: 3 +--- !u!136 &7072444062111318000 +CapsuleCollider: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444062111317993} + m_Material: {fileID: 0} + m_IsTrigger: 0 + m_Enabled: 1 + m_Radius: 0.5 + m_Height: 2.1 + m_Direction: 1 + m_Center: {x: 0, y: 1.02, z: 0} +--- !u!95 &7072444062111318002 +Animator: + serializedVersion: 3 + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444062111317993} + m_Enabled: 1 + m_Avatar: {fileID: 9000000, guid: 860793eb3324391468f1c120a75ec049, type: 3} + m_Controller: {fileID: 9100000, guid: 3ebf60422b6cb1c498ee4cf238072b43, type: 2} + m_CullingMode: 1 + m_UpdateMode: 0 + m_ApplyRootMotion: 0 + m_LinearVelocityBlending: 0 + m_WarningMessage: + m_HasTransformHierarchy: 1 + m_AllowConstantClipSamplingOptimization: 1 + m_KeepAnimatorControllerStateOnDisable: 0 +--- !u!114 &7072444062111318004 +MonoBehaviour: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_GameObject: {fileID: 7072444062111317993} + m_Enabled: 1 + m_EditorHideFlags: 0 + m_Script: {fileID: 11500000, guid: f9c5f555eb7641518e39a97abe893cd8, type: 3} + m_Name: + m_EditorClassIdentifier: + anim: {fileID: 7072444062111318002} + canRotate: 1 diff --git a/Assets/Scenes/dont touch me plz.unity.meta b/Assets/Scenes/AnimationTestScene.unity.meta old mode 100755 new mode 100644 similarity index 74% rename from Assets/Scenes/dont touch me plz.unity.meta rename to Assets/Scenes/AnimationTestScene.unity.meta index b47f91a..2d1f612 --- a/Assets/Scenes/dont touch me plz.unity.meta +++ b/Assets/Scenes/AnimationTestScene.unity.meta @@ -1,5 +1,5 @@ fileFormatVersion: 2 -guid: acbb7f09bd8d02a46bf0793e1aafd61f +guid: b107ebb27f1b8dd40b1e67a2d44d1662 DefaultImporter: externalObjects: {} userData: diff --git a/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity b/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity index 1b0de26..b053540 100644 --- a/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity +++ b/Assets/Scenes/Greatest_map_ever/Greatest_map_ever.unity @@ -1724,102 +1724,24 @@ PrefabInstance: objectReference: {fileID: 0} m_RemovedComponents: [] m_SourcePrefab: {fileID: 100100000, guid: 00815eb34d8b87f42b0215bd3dfd7463, type: 3} ---- !u!1 &1345085340 -GameObject: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - serializedVersion: 6 - m_Component: - - component: {fileID: 1345085341} - - component: {fileID: 1345085344} - - component: {fileID: 1345085343} - - component: {fileID: 1345085342} - m_Layer: 0 - m_Name: navpoint1 - m_TagString: Point - m_Icon: {fileID: 0} - m_NavMeshLayer: 0 - m_StaticEditorFlags: 0 - m_IsActive: 1 ---- !u!4 &1345085341 +--- !u!4 &1345085341 stripped Transform: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} + m_CorrespondingSourceObject: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + m_PrefabInstance: {fileID: 7378066356508352203} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 1345085340} - m_LocalRotation: {x: -0, y: -0, z: -0, w: 1} - m_LocalPosition: {x: 47.446796, y: -5.782543, z: -25.400002} - m_LocalScale: {x: 1.4367, y: 1.4367, z: 1.4367} - m_Children: [] - m_Father: {fileID: 2060099472} - m_RootOrder: 0 - m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} ---- !u!114 &1345085342 +--- !u!114 &1345085342 stripped MonoBehaviour: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} + m_CorrespondingSourceObject: {fileID: 7378066357312323925, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + m_PrefabInstance: {fileID: 7378066356508352203} m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 1345085340} + m_GameObject: {fileID: 0} m_Enabled: 1 m_EditorHideFlags: 0 m_Script: {fileID: 11500000, guid: 6a2d29bfc31a8cf4e831e3bb80720414, type: 3} m_Name: m_EditorClassIdentifier: - navType: 1 - PointId: 0 - DeathAttr: 0 - EnemiesSeen: [] ---- !u!23 &1345085343 -MeshRenderer: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 1345085340} - m_Enabled: 0 - m_CastShadows: 1 - m_ReceiveShadows: 1 - m_DynamicOccludee: 1 - m_MotionVectors: 1 - m_LightProbeUsage: 1 - m_ReflectionProbeUsage: 1 - m_RayTracingMode: 2 - m_RenderingLayerMask: 1 - m_RendererPriority: 0 - m_Materials: - - {fileID: 2100000, guid: 536ae120520e6f245bad32b94a03f6a2, type: 2} - m_StaticBatchInfo: - firstSubMesh: 0 - subMeshCount: 0 - m_StaticBatchRoot: {fileID: 0} - m_ProbeAnchor: {fileID: 0} - m_LightProbeVolumeOverride: {fileID: 0} - m_ScaleInLightmap: 1 - m_ReceiveGI: 1 - m_PreserveUVs: 0 - m_IgnoreNormalsForChartDetection: 0 - m_ImportantGI: 0 - m_StitchLightmapSeams: 1 - m_SelectedEditorRenderState: 3 - m_MinimumChartSize: 4 - m_AutoUVMaxDistance: 0.5 - m_AutoUVMaxAngle: 89 - m_LightmapParameters: {fileID: 0} - m_SortingLayerID: 0 - m_SortingLayer: 0 - m_SortingOrder: 0 ---- !u!33 &1345085344 -MeshFilter: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 1345085340} - m_Mesh: {fileID: 10206, guid: 0000000000000000e000000000000000, type: 0} --- !u!1 &1488699521 GameObject: m_ObjectHideFlags: 0 @@ -2185,6 +2107,7 @@ MonoBehaviour: m_Script: {fileID: 11500000, guid: 593162665e908cf4ea4429f8385dc627, type: 3} m_Name: m_EditorClassIdentifier: + _navPoints: [] --- !u!114 &1858987086 MonoBehaviour: m_ObjectHideFlags: 0 @@ -2826,3 +2749,72 @@ PrefabInstance: objectReference: {fileID: 0} m_RemovedComponents: [] m_SourcePrefab: {fileID: 100100000, guid: 1685c1d9ce4ab174f95c646b1826010b, type: 3} +--- !u!1001 &7378066356508352203 +PrefabInstance: + m_ObjectHideFlags: 0 + serializedVersion: 2 + m_Modification: + m_TransformParent: {fileID: 2060099472} + m_Modifications: + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_RootOrder + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalPosition.x + value: 47.446796 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalPosition.y + value: -5.782543 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalPosition.z + value: -25.400002 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalRotation.w + value: 1 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalRotation.x + value: -0 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalRotation.y + value: -0 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalRotation.z + value: -0 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalEulerAnglesHint.x + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalEulerAnglesHint.y + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323926, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_LocalEulerAnglesHint.z + value: 0 + objectReference: {fileID: 0} + - target: {fileID: 7378066357312323927, guid: f58a14c8ba97dfa4f9f4103bbee852d4, + type: 3} + propertyPath: m_Name + value: navpoint1 + objectReference: {fileID: 0} + m_RemovedComponents: [] + m_SourcePrefab: {fileID: 100100000, guid: f58a14c8ba97dfa4f9f4103bbee852d4, type: 3} diff --git a/Assets/Scenes/First try.unity b/Assets/Scenes/Map#1.unity old mode 100755 new mode 100644 similarity index 100% rename from Assets/Scenes/First try.unity rename to Assets/Scenes/Map#1.unity diff --git a/Assets/Scenes/First try.unity.meta b/Assets/Scenes/Map#1.unity.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Scenes/First try.unity.meta rename to Assets/Scenes/Map#1.unity.meta diff --git a/Assets/Scenes/dont touch me plz.unity b/Assets/Scenes/dont touch me plz.unity deleted file mode 100755 index 3f3b10d..0000000 --- a/Assets/Scenes/dont touch me plz.unity +++ /dev/null @@ -1,460 +0,0 @@ -%YAML 1.1 -%TAG !u! tag:unity3d.com,2011: ---- !u!29 &1 -OcclusionCullingSettings: - m_ObjectHideFlags: 0 - serializedVersion: 2 - m_OcclusionBakeSettings: - smallestOccluder: 5 - smallestHole: 0.25 - backfaceThreshold: 100 - m_SceneGUID: 00000000000000000000000000000000 - m_OcclusionCullingData: {fileID: 0} ---- !u!104 &2 -RenderSettings: - m_ObjectHideFlags: 0 - serializedVersion: 9 - m_Fog: 0 - m_FogColor: {r: 0.5, g: 0.5, b: 0.5, a: 1} - m_FogMode: 3 - m_FogDensity: 0.01 - m_LinearFogStart: 0 - m_LinearFogEnd: 300 - m_AmbientSkyColor: {r: 0.212, g: 0.227, b: 0.259, a: 1} - m_AmbientEquatorColor: {r: 0.114, g: 0.125, b: 0.133, a: 1} - m_AmbientGroundColor: {r: 0.047, g: 0.043, b: 0.035, a: 1} - m_AmbientIntensity: 1 - m_AmbientMode: 0 - m_SubtractiveShadowColor: {r: 0.42, g: 0.478, b: 0.627, a: 1} - m_SkyboxMaterial: {fileID: 10304, guid: 0000000000000000f000000000000000, type: 0} - m_HaloStrength: 0.5 - m_FlareStrength: 1 - m_FlareFadeSpeed: 3 - m_HaloTexture: {fileID: 0} - m_SpotCookie: {fileID: 10001, guid: 0000000000000000e000000000000000, type: 0} - m_DefaultReflectionMode: 0 - m_DefaultReflectionResolution: 128 - m_ReflectionBounces: 1 - m_ReflectionIntensity: 1 - m_CustomReflection: {fileID: 0} - m_Sun: {fileID: 0} - m_IndirectSpecularColor: {r: 0.44657826, g: 0.49641263, b: 0.57481676, a: 1} - m_UseRadianceAmbientProbe: 0 ---- !u!157 &3 -LightmapSettings: - m_ObjectHideFlags: 0 - serializedVersion: 11 - m_GIWorkflowMode: 1 - m_GISettings: - serializedVersion: 2 - m_BounceScale: 1 - m_IndirectOutputScale: 1 - m_AlbedoBoost: 1 - m_EnvironmentLightingMode: 0 - m_EnableBakedLightmaps: 1 - m_EnableRealtimeLightmaps: 0 - m_LightmapEditorSettings: - serializedVersion: 12 - m_Resolution: 2 - m_BakeResolution: 40 - m_AtlasSize: 1024 - m_AO: 0 - m_AOMaxDistance: 1 - m_CompAOExponent: 1 - m_CompAOExponentDirect: 0 - m_ExtractAmbientOcclusion: 0 - m_Padding: 2 - m_LightmapParameters: {fileID: 0} - m_LightmapsBakeMode: 1 - m_TextureCompression: 1 - m_FinalGather: 0 - m_FinalGatherFiltering: 1 - m_FinalGatherRayCount: 256 - m_ReflectionCompression: 2 - m_MixedBakeMode: 2 - m_BakeBackend: 1 - m_PVRSampling: 1 - m_PVRDirectSampleCount: 32 - m_PVRSampleCount: 512 - m_PVRBounces: 2 - m_PVREnvironmentSampleCount: 256 - m_PVREnvironmentReferencePointCount: 2048 - m_PVRFilteringMode: 1 - m_PVRDenoiserTypeDirect: 1 - m_PVRDenoiserTypeIndirect: 1 - m_PVRDenoiserTypeAO: 1 - m_PVRFilterTypeDirect: 0 - m_PVRFilterTypeIndirect: 0 - m_PVRFilterTypeAO: 0 - m_PVREnvironmentMIS: 1 - m_PVRCulling: 1 - m_PVRFilteringGaussRadiusDirect: 1 - m_PVRFilteringGaussRadiusIndirect: 5 - m_PVRFilteringGaussRadiusAO: 2 - m_PVRFilteringAtrousPositionSigmaDirect: 0.5 - m_PVRFilteringAtrousPositionSigmaIndirect: 2 - m_PVRFilteringAtrousPositionSigmaAO: 1 - m_ExportTrainingData: 0 - m_TrainingDataDestination: TrainingData - m_LightProbeSampleCountMultiplier: 4 - m_LightingDataAsset: {fileID: 0} - m_UseShadowmask: 1 ---- !u!196 &4 -NavMeshSettings: - serializedVersion: 2 - m_ObjectHideFlags: 0 - m_BuildSettings: - serializedVersion: 2 - agentTypeID: 0 - agentRadius: 0.5 - agentHeight: 2 - agentSlope: 45 - agentClimb: 0.4 - ledgeDropHeight: 0 - maxJumpAcrossDistance: 0 - minRegionArea: 2 - manualCellSize: 0 - cellSize: 0.16666667 - manualTileSize: 0 - tileSize: 256 - accuratePlacement: 0 - debug: - m_Flags: 0 - m_NavMeshData: {fileID: 0} ---- !u!1 &254804098 -GameObject: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - serializedVersion: 6 - m_Component: - - component: {fileID: 254804102} - - component: {fileID: 254804101} - - component: {fileID: 254804100} - - component: {fileID: 254804099} - m_Layer: 0 - m_Name: Cube - m_TagString: Untagged - m_Icon: {fileID: 0} - m_NavMeshLayer: 0 - m_StaticEditorFlags: 0 - m_IsActive: 1 ---- !u!65 &254804099 -BoxCollider: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 254804098} - m_Material: {fileID: 0} - m_IsTrigger: 0 - m_Enabled: 1 - serializedVersion: 2 - m_Size: {x: 1, y: 1, z: 1} - m_Center: {x: 0, y: 0, z: 0} ---- !u!23 &254804100 -MeshRenderer: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 254804098} - m_Enabled: 1 - m_CastShadows: 1 - m_ReceiveShadows: 1 - m_DynamicOccludee: 1 - m_MotionVectors: 1 - m_LightProbeUsage: 1 - m_ReflectionProbeUsage: 1 - m_RayTracingMode: 2 - m_RenderingLayerMask: 1 - m_RendererPriority: 0 - m_Materials: - - {fileID: 2100000, guid: d911963c0a55d5243b20ebed4dddc9fd, type: 2} - m_StaticBatchInfo: - firstSubMesh: 0 - subMeshCount: 0 - m_StaticBatchRoot: {fileID: 0} - m_ProbeAnchor: {fileID: 0} - m_LightProbeVolumeOverride: {fileID: 0} - m_ScaleInLightmap: 1 - m_ReceiveGI: 1 - m_PreserveUVs: 0 - m_IgnoreNormalsForChartDetection: 0 - m_ImportantGI: 0 - m_StitchLightmapSeams: 1 - m_SelectedEditorRenderState: 3 - m_MinimumChartSize: 4 - m_AutoUVMaxDistance: 0.5 - m_AutoUVMaxAngle: 89 - m_LightmapParameters: {fileID: 0} - m_SortingLayerID: 0 - m_SortingLayer: 0 - m_SortingOrder: 0 ---- !u!33 &254804101 -MeshFilter: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 254804098} - m_Mesh: {fileID: 10202, guid: 0000000000000000e000000000000000, type: 0} ---- !u!4 &254804102 -Transform: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 254804098} - m_LocalRotation: {x: 0, y: 0, z: 0, w: 1} - m_LocalPosition: {x: 49.153797, y: 9.146889, z: 40.07178} - m_LocalScale: {x: 100, y: 1, z: 100} - m_Children: [] - m_Father: {fileID: 0} - m_RootOrder: 2 - m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} ---- !u!1 &676942498 -GameObject: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - serializedVersion: 6 - m_Component: - - component: {fileID: 676942500} - - component: {fileID: 676942499} - m_Layer: 0 - m_Name: Directional Light - m_TagString: Untagged - m_Icon: {fileID: 0} - m_NavMeshLayer: 0 - m_StaticEditorFlags: 0 - m_IsActive: 1 ---- !u!108 &676942499 -Light: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 676942498} - m_Enabled: 1 - serializedVersion: 10 - m_Type: 1 - m_Shape: 0 - m_Color: {r: 1, g: 0.95686275, b: 0.8392157, a: 1} - m_Intensity: 1 - m_Range: 10 - m_SpotAngle: 30 - m_InnerSpotAngle: 21.80208 - m_CookieSize: 10 - m_Shadows: - m_Type: 2 - m_Resolution: -1 - m_CustomResolution: -1 - m_Strength: 1 - m_Bias: 0.05 - m_NormalBias: 0.4 - m_NearPlane: 0.2 - m_CullingMatrixOverride: - e00: 1 - e01: 0 - e02: 0 - e03: 0 - e10: 0 - e11: 1 - e12: 0 - e13: 0 - e20: 0 - e21: 0 - e22: 1 - e23: 0 - e30: 0 - e31: 0 - e32: 0 - e33: 1 - m_UseCullingMatrixOverride: 0 - m_Cookie: {fileID: 0} - m_DrawHalo: 0 - m_Flare: {fileID: 0} - m_RenderMode: 0 - m_CullingMask: - serializedVersion: 2 - m_Bits: 4294967295 - m_RenderingLayerMask: 1 - m_Lightmapping: 4 - m_LightShadowCasterMode: 0 - m_AreaSize: {x: 1, y: 1} - m_BounceIntensity: 1 - m_ColorTemperature: 6570 - m_UseColorTemperature: 0 - m_BoundingSphereOverride: {x: 0, y: 0, z: 0, w: 0} - m_UseBoundingSphereOverride: 0 - m_ShadowRadius: 0 - m_ShadowAngle: 0 ---- !u!4 &676942500 -Transform: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 676942498} - m_LocalRotation: {x: 0.40821788, y: -0.23456968, z: 0.10938163, w: 0.8754261} - m_LocalPosition: {x: 0, y: 3, z: 0} - m_LocalScale: {x: 1, y: 1, z: 1} - m_Children: [] - m_Father: {fileID: 0} - m_RootOrder: 1 - m_LocalEulerAnglesHint: {x: 50, y: -30, z: 0} ---- !u!1 &1991868003 -GameObject: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - serializedVersion: 6 - m_Component: - - component: {fileID: 1991868006} - - component: {fileID: 1991868005} - - component: {fileID: 1991868004} - m_Layer: 0 - m_Name: Main Camera - m_TagString: MainCamera - m_Icon: {fileID: 0} - m_NavMeshLayer: 0 - m_StaticEditorFlags: 0 - m_IsActive: 1 ---- !u!81 &1991868004 -AudioListener: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 1991868003} - m_Enabled: 1 ---- !u!20 &1991868005 -Camera: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 1991868003} - m_Enabled: 1 - serializedVersion: 2 - m_ClearFlags: 1 - m_BackGroundColor: {r: 0.19215687, g: 0.3019608, b: 0.4745098, a: 0} - m_projectionMatrixMode: 1 - m_GateFitMode: 2 - m_FOVAxisMode: 0 - m_SensorSize: {x: 36, y: 24} - m_LensShift: {x: 0, y: 0} - m_FocalLength: 50 - m_NormalizedViewPortRect: - serializedVersion: 2 - x: 0 - y: 0 - width: 1 - height: 1 - near clip plane: 0.3 - far clip plane: 1000 - field of view: 60 - orthographic: 0 - orthographic size: 5 - m_Depth: -1 - m_CullingMask: - serializedVersion: 2 - m_Bits: 4294967295 - m_RenderingPath: -1 - m_TargetTexture: {fileID: 0} - m_TargetDisplay: 0 - m_TargetEye: 3 - m_HDR: 1 - m_AllowMSAA: 1 - m_AllowDynamicResolution: 0 - m_ForceIntoRT: 0 - m_OcclusionCulling: 1 - m_StereoConvergence: 10 - m_StereoSeparation: 0.022 ---- !u!4 &1991868006 -Transform: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_GameObject: {fileID: 1991868003} - m_LocalRotation: {x: 0, y: 0, z: 0, w: 1} - m_LocalPosition: {x: 0, y: 1, z: -10} - m_LocalScale: {x: 1, y: 1, z: 1} - m_Children: [] - m_Father: {fileID: 0} - m_RootOrder: 0 - m_LocalEulerAnglesHint: {x: 0, y: 0, z: 0} ---- !u!1001 &8510909887784509496 -PrefabInstance: - m_ObjectHideFlags: 0 - serializedVersion: 2 - m_Modification: - m_TransformParent: {fileID: 0} - m_Modifications: - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_RootOrder - value: 3 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalPosition.x - value: 32.94 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalPosition.y - value: 10.9 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalPosition.z - value: 46.67 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalRotation.w - value: 0.7071068 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalRotation.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalRotation.y - value: 0.7071068 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalRotation.z - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalEulerAnglesHint.x - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalEulerAnglesHint.y - value: 90 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480103, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_LocalEulerAnglesHint.z - value: 0 - objectReference: {fileID: 0} - - target: {fileID: 5245491127989480125, guid: 80f6c1c85e5daed4c96c70205ed5503d, - type: 3} - propertyPath: m_Name - value: Player - objectReference: {fileID: 0} - m_RemovedComponents: [] - m_SourcePrefab: {fileID: 100100000, guid: 80f6c1c85e5daed4c96c70205ed5503d, type: 3} diff --git a/Assets/Scripts/Animators.meta b/Assets/Scripts/Animators.meta new file mode 100644 index 0000000..ad7ca65 --- /dev/null +++ b/Assets/Scripts/Animators.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: 0cf97b021cee45eb8f7d402f16955139 +timeCreated: 1652022637 \ No newline at end of file diff --git a/Assets/Scripts/Animators/Kirill Animator.meta b/Assets/Scripts/Animators/Kirill Animator.meta new file mode 100644 index 0000000..6ef1fdf --- /dev/null +++ b/Assets/Scripts/Animators/Kirill Animator.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: 97bb456bbc4248378002eefeb52be6c3 +timeCreated: 1652022665 \ No newline at end of file diff --git a/Assets/Scripts/Character/scr_CharacterController.cs b/Assets/Scripts/Animators/Kirill Animator/CustomCharacterController.cs old mode 100755 new mode 100644 similarity index 95% rename from Assets/Scripts/Character/scr_CharacterController.cs rename to Assets/Scripts/Animators/Kirill Animator/CustomCharacterController.cs index e4f847d..9276565 --- a/Assets/Scripts/Character/scr_CharacterController.cs +++ b/Assets/Scripts/Animators/Kirill Animator/CustomCharacterController.cs @@ -1,255 +1,253 @@ -using UnityEngine; - -using static scr_Models; - -public class scr_CharacterController : MonoBehaviour -{ - - private CharacterController characterController; - private DefaultInput defaultInput; - private Vector2 input_Movement; - [HideInInspector] - public Vector2 input_View; - - private Vector3 newCameraRotation; - private Vector3 newCharacterRotation; - - [Header("References")] - public Transform cameraHolder; - public Transform feetTransform; - - [Header("Settings")] - public PlayerSettingsModel playerSettings; - - public float ViewClampYMin = -70; - public float ViewClampYMax = 80; - public LayerMask playerMask; - - [Header("Gravity")] - public float gravityAmount; - public float gravityMin; - private float playerGravity; - - public Vector3 jumpingForce; - private Vector3 jumpingForceVelocity; - - [Header("Stance")] - public PlayerStance playerStance; - public float playerStanceSmoothing; - public CharacterStance playerStandStance; - public CharacterStance playerCrouchStance; - public CharacterStance playerProneStance; - private float stanceCheckErrorMargin = 0.05f; - - private float cameraHeight; - private float cameraHeightVelocity; - - private bool isSprinting; - - private Vector3 newMovementSpeed; - private Vector3 newMovementSpeedVelocity; - - [Header("Weapon")] public scr_WeaponController currentWeapon; - private void Awake() - { - defaultInput = new DefaultInput(); - - defaultInput.Character.Movement.performed += e => input_Movement = e.ReadValue(); - defaultInput.Character.View.performed += e => input_View = e.ReadValue(); - defaultInput.Character.Jump.performed += e => Jump(); - - defaultInput.Character.Crouch.performed += e => Crouch(); - defaultInput.Character.Prone.performed += e => Prone(); - - defaultInput.Character.Sprint.performed += e => ToggleSprint(); - defaultInput.Character.SprintReleased.performed += e => StopSprint(); - - defaultInput.Enable(); - - newCameraRotation = cameraHolder.localRotation.eulerAngles; - newCharacterRotation = transform.localRotation.eulerAngles; - characterController = GetComponent(); - - cameraHeight = cameraHolder.localPosition.y; - - if (currentWeapon) - { - currentWeapon.Initialise(this); - } - - } - - private void Update() - { - CalculateView(); - CalculateMovement(); - CalculateJump(); - CalculateCameraHeight(); - } - - private void CalculateView() - { - newCharacterRotation.y += playerSettings.ViewXSensetivity * (playerSettings.ViewXInverted ? -input_View.x : input_View.x) * Time.deltaTime; - transform.localRotation = Quaternion.Euler(newCharacterRotation); - - newCameraRotation.x += playerSettings.ViewYSensetivity * (playerSettings.ViewYInverted ? input_View.y : -input_View.y) * Time.deltaTime; - newCameraRotation.x = Mathf.Clamp(newCameraRotation.x, ViewClampYMin, ViewClampYMax); - - cameraHolder.localRotation = Quaternion.Euler(newCameraRotation); - } - - private void CalculateMovement() - { - if (input_Movement.y <= 0.2f) - { - isSprinting = false; - } - - var verticalSpeed = playerSettings.WalkingForwardSpeed; - var horizontalSpeed = playerSettings.WalkingStrafeSpeed; - - if (isSprinting) - { - verticalSpeed = playerSettings.RunningForwardSpeed; - horizontalSpeed = playerSettings.RunningStrafeSpeed; - } - - // Effectors - if (!characterController.isGrounded) - { - playerSettings.SpeedEffector = playerSettings.FallingSpeedEffector; - } - else if (playerStance == PlayerStance.Crouch) - { - playerSettings.SpeedEffector = playerSettings.CrouchSpeedEffector; - } - else if (playerStance == PlayerStance.Prone) - { - playerSettings.SpeedEffector = playerSettings.ProneSpeedEffector; - } - else - { - playerSettings.SpeedEffector = 1; - } - - verticalSpeed *= playerSettings.SpeedEffector; - horizontalSpeed *= playerSettings.SpeedEffector; - - newMovementSpeed = Vector3.SmoothDamp(newMovementSpeed, - new Vector3(horizontalSpeed * input_Movement.x * Time.deltaTime, - 0, verticalSpeed * input_Movement.y * Time.deltaTime), - ref newMovementSpeedVelocity, characterController.isGrounded ? playerSettings.MovementSmoothing : playerSettings.FallingSmoothing); - - var MovementSpeed = transform.TransformDirection(newMovementSpeed); - - if (playerGravity > gravityMin) - { - playerGravity -= gravityAmount * Time.deltaTime; - } - - if (playerGravity < -0.1f && characterController.isGrounded) - { - playerGravity = -0.1f; - } - - MovementSpeed.y += playerGravity; - MovementSpeed += jumpingForce * Time.deltaTime; - - characterController.Move(MovementSpeed); - } - - private void CalculateJump() - { - jumpingForce = Vector3.SmoothDamp(jumpingForce, Vector3.zero, ref jumpingForceVelocity, playerSettings.JumpingFalloff); - } - - private void CalculateCameraHeight() - { - var stanceHeight = playerStandStance.CameraHeight; - - if (playerStance == PlayerStance.Crouch) - { - stanceHeight = playerCrouchStance.CameraHeight; - } - else if (playerStance == PlayerStance.Prone) - { - stanceHeight = playerProneStance.CameraHeight; - } - - cameraHeight = Mathf.SmoothDamp(cameraHolder.localPosition.y, stanceHeight, ref cameraHeightVelocity, playerStanceSmoothing); - - cameraHolder.localPosition = new Vector3(cameraHolder.localPosition.x, cameraHeight, cameraHolder.localPosition.z); - } - private void Jump() - { - if (!characterController.isGrounded || playerStance == PlayerStance.Prone) - { - return; - } - - if (playerStance == PlayerStance.Crouch) - { - if (StanceCheck(playerStandStance.StanceCollider.height)) - { - return; - } - playerStance = PlayerStance.Stand; - return; - } - - // Jump - jumpingForce = Vector3.up * playerSettings.JumpingHeight; - playerGravity = 0; - } - - private void Crouch() - { - if (playerStance == PlayerStance.Crouch) - { - if (StanceCheck(playerStandStance.StanceCollider.height)) - { - return; - } - playerStance = PlayerStance.Stand; - return; - } - if (StanceCheck(playerCrouchStance.StanceCollider.height)) - { - return; - } - playerStance = PlayerStance.Crouch; - } - - private void Prone() - { - playerStance = PlayerStance.Prone; - } - - private bool StanceCheck(float stanceCheckheight) - { - var start = new Vector3(feetTransform.position.x, feetTransform.position.y + characterController.radius + stanceCheckErrorMargin, feetTransform.position.z); - var end = new Vector3(feetTransform.position.x, feetTransform.position.y - characterController.radius - stanceCheckErrorMargin + stanceCheckheight, feetTransform.position.z); - - - return Physics.CheckCapsule(start, end, characterController.radius, playerMask); - } - - private void ToggleSprint() - { - if (input_Movement.y <= 0.2f) - { - isSprinting = false; - return; - } - isSprinting = !isSprinting; - } - - private void StopSprint() - { - if (playerSettings.SprintingHold) - { - isSprinting = false; - } - } - -} +using UnityEngine; + +public class scr_CharacterController : MonoBehaviour +{ + private CharacterController characterController; + private DefaultInput defaultInput; + private Vector2 input_Movement; + [HideInInspector] + public Vector2 input_View; + + private Vector3 newCameraRotation; + private Vector3 newCharacterRotation; + + [Header("References")] + public Transform cameraHolder; + public Transform feetTransform; + + [Header("Settings")] + public PlayerSettingsModel playerSettings; + + public float ViewClampYMin = -70; + public float ViewClampYMax = 80; + public LayerMask playerMask; + + [Header("Gravity")] + public float gravityAmount; + public float gravityMin; + private float playerGravity; + + public Vector3 jumpingForce; + private Vector3 jumpingForceVelocity; + + [Header("Stance")] + public PlayerStance playerStance; + public float playerStanceSmoothing; + public CharacterStance playerStandStance; + public CharacterStance playerCrouchStance; + public CharacterStance playerProneStance; + private float stanceCheckErrorMargin = 0.05f; + + private float cameraHeight; + private float cameraHeightVelocity; + + private bool isSprinting; + + private Vector3 newMovementSpeed; + private Vector3 newMovementSpeedVelocity; + + [Header("Weapon")] + public scr_WeaponController currentWeapon; + private void Awake() + { + defaultInput = new DefaultInput(); + + defaultInput.Character.Movement.performed += e => input_Movement = e.ReadValue(); + defaultInput.Character.View.performed += e => input_View = e.ReadValue(); + defaultInput.Character.Jump.performed += e => Jump(); + + defaultInput.Character.Crouch.performed += e => Crouch(); + defaultInput.Character.Prone.performed += e => Prone(); + + defaultInput.Character.Sprint.performed += e => ToggleSprint(); + defaultInput.Character.SprintReleased.performed += e => StopSprint(); + + defaultInput.Enable(); + + newCameraRotation = cameraHolder.localRotation.eulerAngles; + newCharacterRotation = transform.localRotation.eulerAngles; + characterController = GetComponent(); + + cameraHeight = cameraHolder.localPosition.y; + + if (currentWeapon) + { + currentWeapon.Initialise(this); + } + + } + + private void Update() + { + CalculateView(); + CalculateMovement(); + CalculateJump(); + CalculateCameraHeight(); + } + + private void CalculateView() + { + newCharacterRotation.y += playerSettings.ViewXSensetivity * (playerSettings.ViewXInverted ? -input_View.x : input_View.x) * Time.deltaTime; + transform.localRotation = Quaternion.Euler(newCharacterRotation); + + newCameraRotation.x += playerSettings.ViewYSensetivity * (playerSettings.ViewYInverted ? input_View.y : -input_View.y) * Time.deltaTime; + newCameraRotation.x = Mathf.Clamp(newCameraRotation.x, ViewClampYMin, ViewClampYMax); + + cameraHolder.localRotation = Quaternion.Euler(newCameraRotation); + } + + private void CalculateMovement() + { + if (input_Movement.y <= 0.2f) + { + isSprinting = false; + } + + var verticalSpeed = playerSettings.WalkingForwardSpeed; + var horizontalSpeed = playerSettings.WalkingStrafeSpeed; + + if (isSprinting) + { + verticalSpeed = playerSettings.RunningForwardSpeed; + horizontalSpeed = playerSettings.RunningStrafeSpeed; + } + + // Effectors + if (!characterController.isGrounded) + { + playerSettings.SpeedEffector = playerSettings.FallingSpeedEffector; + } + else if (playerStance == PlayerStance.Crouch) + { + playerSettings.SpeedEffector = playerSettings.CrouchSpeedEffector; + } + else if (playerStance == PlayerStance.Prone) + { + playerSettings.SpeedEffector = playerSettings.ProneSpeedEffector; + } + else + { + playerSettings.SpeedEffector = 1; + } + + verticalSpeed *= playerSettings.SpeedEffector; + horizontalSpeed *= playerSettings.SpeedEffector; + + newMovementSpeed = Vector3.SmoothDamp(newMovementSpeed, + new Vector3(horizontalSpeed * input_Movement.x * Time.deltaTime, + 0, verticalSpeed * input_Movement.y * Time.deltaTime), + ref newMovementSpeedVelocity, characterController.isGrounded ? playerSettings.MovementSmoothing : playerSettings.FallingSmoothing); + + var MovementSpeed = transform.TransformDirection(newMovementSpeed); + + if (playerGravity > gravityMin) + { + playerGravity -= gravityAmount * Time.deltaTime; + } + + if (playerGravity < -0.1f && characterController.isGrounded) + { + playerGravity = -0.1f; + } + + MovementSpeed.y += playerGravity; + MovementSpeed += jumpingForce * Time.deltaTime; + + characterController.Move(MovementSpeed); + } + + private void CalculateJump() + { + jumpingForce = Vector3.SmoothDamp(jumpingForce, Vector3.zero, ref jumpingForceVelocity, playerSettings.JumpingFalloff); + } + + private void CalculateCameraHeight() + { + var stanceHeight = playerStandStance.CameraHeight; + + if (playerStance == PlayerStance.Crouch) + { + stanceHeight = playerCrouchStance.CameraHeight; + } + else if (playerStance == PlayerStance.Prone) + { + stanceHeight = playerProneStance.CameraHeight; + } + + cameraHeight = Mathf.SmoothDamp(cameraHolder.localPosition.y, stanceHeight, ref cameraHeightVelocity, playerStanceSmoothing); + + cameraHolder.localPosition = new Vector3(cameraHolder.localPosition.x, cameraHeight, cameraHolder.localPosition.z); + } + private void Jump() + { + if (!characterController.isGrounded || playerStance == PlayerStance.Prone) + { + return; + } + + if (playerStance == PlayerStance.Crouch) + { + if (StanceCheck(playerStandStance.StanceCollider.height)) + { + return; + } + playerStance = PlayerStance.Stand; + return; + } + + // Jump + jumpingForce = Vector3.up * playerSettings.JumpingHeight; + playerGravity = 0; + } + + private void Crouch() + { + if (playerStance == PlayerStance.Crouch) + { + if (StanceCheck(playerStandStance.StanceCollider.height)) + { + return; + } + playerStance = PlayerStance.Stand; + return; + } + if (StanceCheck(playerCrouchStance.StanceCollider.height)) + { + return; + } + playerStance = PlayerStance.Crouch; + } + + private void Prone() + { + playerStance = PlayerStance.Prone; + } + + private bool StanceCheck(float stanceCheckheight) + { + var start = new Vector3(feetTransform.position.x, feetTransform.position.y + characterController.radius + stanceCheckErrorMargin, feetTransform.position.z); + var end = new Vector3(feetTransform.position.x, feetTransform.position.y - characterController.radius - stanceCheckErrorMargin + stanceCheckheight, feetTransform.position.z); + + + return Physics.CheckCapsule(start, end, characterController.radius, playerMask); + } + + private void ToggleSprint() + { + if (input_Movement.y <= 0.2f) + { + isSprinting = false; + return; + } + isSprinting = !isSprinting; + } + + private void StopSprint() + { + if (playerSettings.SprintingHold) + { + isSprinting = false; + } + } + +} diff --git a/Assets/Scripts/Character/scr_CharacterController.cs.meta b/Assets/Scripts/Animators/Kirill Animator/CustomCharacterController.cs.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Scripts/Character/scr_CharacterController.cs.meta rename to Assets/Scripts/Animators/Kirill Animator/CustomCharacterController.cs.meta diff --git a/Assets/Scripts/Animators/Kirill Animator/Models.cs b/Assets/Scripts/Animators/Kirill Animator/Models.cs new file mode 100644 index 0000000..5e4ec9d --- /dev/null +++ b/Assets/Scripts/Animators/Kirill Animator/Models.cs @@ -0,0 +1,67 @@ +using System; +using UnityEngine; + + +public enum PlayerStance +{ + Stand, + Crouch, + Prone +} + +[Serializable] +public class PlayerSettingsModel +{ + [Header("View Settings")] + public float ViewXSensetivity; + public float ViewYSensetivity; + + public bool ViewXInverted; + public bool ViewYInverted; + + [Header("Movement Settings")] + public bool SprintingHold; + public float MovementSmoothing; + + [Header("Movement - Running")] + public float RunningForwardSpeed; + public float RunningStrafeSpeed; + + [Header("Movement - Walking")] + public float WalkingForwardSpeed; + public float WalkingBackwardSpeed; + public float WalkingStrafeSpeed; + + [Header("Jumping")] + public float JumpingHeight; + public float JumpingFalloff; + public float FallingSmoothing; + + [Header("Speed Effectors")] + public float SpeedEffector = 1; + public float CrouchSpeedEffector; + public float ProneSpeedEffector; + public float FallingSpeedEffector; +} + + +[Serializable] +public class CharacterStance +{ + public float CameraHeight; + public CapsuleCollider StanceCollider; +} + + +[Serializable] +public class WeaponSettingsModel +{ + [Header("Sway")] + public float SwayAmount; + public bool SwayYInverted; + public bool SwayXInverted; + public float SwaySmoothing; + public float SwayResetSmoothing; + public float SwayClampX; + public float SwayClampY; +} \ No newline at end of file diff --git a/Assets/Scripts/Character/scr_Models.cs.meta b/Assets/Scripts/Animators/Kirill Animator/Models.cs.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Scripts/Character/scr_Models.cs.meta rename to Assets/Scripts/Animators/Kirill Animator/Models.cs.meta diff --git a/Assets/Scripts/Animators/Leonid Animator.meta b/Assets/Scripts/Animators/Leonid Animator.meta new file mode 100644 index 0000000..e6c9ea6 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: 2cc90c830ef641a1a18e8e21dc97dec0 +timeCreated: 1652022681 \ No newline at end of file diff --git a/Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs b/Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs new file mode 100644 index 0000000..6821141 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs @@ -0,0 +1,94 @@ +using Unity.Mathematics; +using UnityEngine; + +namespace Animators.Leonid_Animator +{ + public class AnimatorHandler : MonoBehaviour + { + public Animator anim; + public bool canRotate; + + private int _horizontal; + private int _vertical; + private bool _isCrouching = false; + private bool _isJumping; + + private int _crouch; + private int _jump; + private int _fired; + + public void Initialize() + { + anim = GetComponent(); + _vertical = Animator.StringToHash(nameof(_vertical)); + _horizontal = Animator.StringToHash(nameof(_horizontal)); + _crouch = Animator.StringToHash(nameof(_crouch)); + _jump = Animator.StringToHash(nameof(_jump)); + _fired = Animator.StringToHash(nameof(_fired)); + } + + public void UpdateAnimatorValues(float verticalMovement, float horizontalMovement, + bool pressedJumped, bool pressedCrouching, bool firePressed) + { + #region Vertical Movement + + var vertical = 0f; + if (verticalMovement > 0 && verticalMovement < 0.55) + vertical = 0.5f; + else if (verticalMovement > 0.55) + vertical = 1; + else if (verticalMovement < 0 && verticalMovement > -0.55) + { + vertical = -0.5f; + } + else if (verticalMovement < -0.55) + { + vertical = -1; + } + else + { + vertical = 0; + } + #endregion + + #region Vertical Movement + + var horizontal = 0f; + if (horizontalMovement > 0 && horizontalMovement < 0.55) + horizontal = 0.5f; + else if (horizontalMovement > 0.55) + horizontal = 1; + else if (horizontalMovement < 0 && horizontalMovement > -0.55) + { + horizontal = -0.5f; + } + else if (horizontalMovement < -0.55) + { + horizontal = -1; + } + else + { + horizontal = 0; + } + #endregion + + anim.SetFloat(_horizontal, horizontal, 0.1f, Time.deltaTime); + anim.SetFloat(_vertical, vertical, 0.1f, Time.deltaTime); + + if (pressedCrouching == true) + { + _isCrouching = !_isCrouching; + if (_isCrouching == true) + transform.Rotate(Vector3.up, 45); + else + { + transform.Rotate(Vector3.up, -45); + } + anim.SetBool(_crouch, _isCrouching); + } + + anim.SetBool(_jump, pressedJumped); + anim.SetBool(_fired, firePressed); + } + } +} \ No newline at end of file diff --git a/Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs.meta b/Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs.meta new file mode 100644 index 0000000..a475077 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/AnimatorHandler.cs.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: f9c5f555eb7641518e39a97abe893cd8 +timeCreated: 1652031215 \ No newline at end of file diff --git a/Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller b/Assets/Scripts/Animators/Leonid Animator/CharacterAnimator.controller old mode 100755 new mode 100644 similarity index 56% rename from Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller rename to Assets/Scripts/Animators/Leonid Animator/CharacterAnimator.controller index a2a4e1d..12a24c7 --- a/Assets/SciFiWarriorPBRHPPolyart/Animators/SciFiWarrior.controller +++ b/Assets/Scripts/Animators/Leonid Animator/CharacterAnimator.controller @@ -1,358 +1,109 @@ %YAML 1.1 %TAG !u! tag:unity3d.com,2011: ---- !u!91 &9100000 -AnimatorController: - m_ObjectHideFlags: 0 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: SciFiWarrior - serializedVersion: 5 - m_AnimatorParameters: [] - m_AnimatorLayers: - - serializedVersion: 5 - m_Name: Base Layer - m_StateMachine: {fileID: 1107476691846109156} - m_Mask: {fileID: 0} - m_Motions: [] - m_Behaviours: [] - m_BlendingMode: 0 - m_SyncedLayerIndex: -1 - m_DefaultWeight: 0 - m_IKPass: 0 - m_SyncedLayerAffectsTiming: 0 - m_Controller: {fileID: 9100000} ---- !u!1101 &1101005713632051030 +--- !u!1101 &-8614502741554326989 AnimatorStateTransition: m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} m_Name: - m_Conditions: [] + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102885244088947746} + m_DstState: {fileID: -5023192667791512651} m_Solo: 0 m_Mute: 0 m_IsExit: 0 serializedVersion: 3 - m_TransitionDuration: 0.25 + m_TransitionDuration: 0.46723264 + m_TransitionOffset: 0.21265899 + m_ExitTime: 0.57224524 + m_HasExitTime: 0 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1107 &-8265500127550764659 +AnimatorStateMachine: + serializedVersion: 6 + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Upper + m_ChildStates: + - serializedVersion: 1 + m_State: {fileID: -4600760231423422918} + m_Position: {x: 340, y: 140, z: 0} + - serializedVersion: 1 + m_State: {fileID: 7739110899394721029} + m_Position: {x: 510, y: -10, z: 0} + - serializedVersion: 1 + m_State: {fileID: -4187437994059944167} + m_Position: {x: 400, y: 290, z: 0} + - serializedVersion: 1 + m_State: {fileID: -5023192667791512651} + m_Position: {x: 830, y: 130, z: 0} + m_ChildStateMachines: [] + m_AnyStateTransitions: [] + m_EntryTransitions: [] + m_StateMachineTransitions: {} + m_StateMachineBehaviours: [] + m_AnyStatePosition: {x: 50, y: 20, z: 0} + m_EntryPosition: {x: 50, y: 120, z: 0} + m_ExitPosition: {x: 1150, y: 160, z: 0} + m_ParentStateMachinePosition: {x: 800, y: 20, z: 0} + m_DefaultState: {fileID: -4600760231423422918} +--- !u!1101 &-6757359955429936644 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Jump to Run + m_Conditions: [] + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 8354844821256608690} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.20300466 m_TransitionOffset: 0 - m_ExitTime: 4 + m_ExitTime: 0.7916667 m_HasExitTime: 1 m_HasFixedDuration: 1 m_InterruptionSource: 0 m_OrderedInterruption: 1 m_CanTransitionToSelf: 1 ---- !u!1101 &1101018267375076786 +--- !u!1101 &-6673604382440492192 AnimatorStateTransition: m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} m_Name: - m_Conditions: [] + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _fired + m_EventTreshold: 0 m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102942199553133880} + m_DstState: {fileID: -4187437994059944167} m_Solo: 0 m_Mute: 0 m_IsExit: 0 serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 + m_TransitionDuration: 1.2815269 + m_TransitionOffset: 0.04597427 + m_ExitTime: 0.009130422 + m_HasExitTime: 0 m_HasFixedDuration: 1 m_InterruptionSource: 0 m_OrderedInterruption: 1 m_CanTransitionToSelf: 1 ---- !u!1101 &1101029586256411172 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102525727787806748} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101085741205891168 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102100670134039114} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101213140064716440 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102901032506865742} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 6 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101350354845918052 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102106218193099826} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 6 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101363450263446180 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102243588419055754} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 3 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101376048922213750 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102949299928912516} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101404391616814046 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102379806414260514} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101429387431206348 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102277268960662982} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101567676818271368 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102238716766734490} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101616875765159898 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102642198119822076} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 6 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101624051492446352 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102361151154106906} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 4 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101871786086276220 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102625340767904258} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 10 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1101 &1101995184118313822 -AnimatorStateTransition: - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: - m_Conditions: [] - m_DstStateMachine: {fileID: 0} - m_DstState: {fileID: 1102088155696507906} - m_Solo: 0 - m_Mute: 0 - m_IsExit: 0 - serializedVersion: 3 - m_TransitionDuration: 0.25 - m_TransitionOffset: 0 - m_ExitTime: 6 - m_HasExitTime: 1 - m_HasFixedDuration: 1 - m_InterruptionSource: 0 - m_OrderedInterruption: 1 - m_CanTransitionToSelf: 1 ---- !u!1102 &1102088155696507906 +--- !u!1102 &-5914497066343941395 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 @@ -363,7 +114,8 @@ AnimatorState: m_Speed: 1 m_CycleOffset: 0 m_Transitions: - - {fileID: 1101429387431206348} + - {fileID: 5168308916736617153} + - {fileID: 9141976730198879995} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -379,115 +131,32 @@ AnimatorState: m_MirrorParameter: m_CycleOffsetParameter: m_TimeParameter: ---- !u!1102 &1102100670134039114 -AnimatorState: - serializedVersion: 6 +--- !u!1101 &-5341886129914063569 +AnimatorStateTransition: m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: Idle_Shoot_Ar - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101005713632051030} - m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 1d4365e1541bb6949a273318862b72d3, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102106218193099826 -AnimatorState: - serializedVersion: 6 - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: Shoot_Autoshot_AR - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101871786086276220} - m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 76a61eccc11250c4da8bd7f1d87fa2fe, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102238716766734490 -AnimatorState: - serializedVersion: 6 - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: Idle_gunMiddle_AR - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101085741205891168} - m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 107649059ea401b4e9c5c20f21e99a55, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102243588419055754 -AnimatorState: - serializedVersion: 6 - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: WalkFront_Shoot_AR - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101029586256411172} - m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: e398e3aa9cdfcdc4ebd01a06e4c544ae, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102277268960662982 + m_Name: + m_Conditions: + - m_ConditionMode: 2 + m_ConditionEvent: _jump + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: -4600760231423422918} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.7916667 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1102 &-5023192667791512651 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 @@ -497,7 +166,8 @@ AnimatorState: m_Name: Die m_Speed: 1 m_CycleOffset: 0 - m_Transitions: [] + m_Transitions: + - {fileID: 1131199853383832992} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -513,18 +183,70 @@ AnimatorState: m_MirrorParameter: m_CycleOffsetParameter: m_TimeParameter: ---- !u!1102 &1102361151154106906 +--- !u!1101 &-4911913766117122026 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 2 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 0} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 1 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.765625 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &-4865886577319040672 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Run to Jump + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _jump + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: -5914497066343941395} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.75 + m_HasExitTime: 0 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1102 &-4600760231423422918 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: WalkRight_Shoot_AR + m_Name: Idle m_Speed: 1 m_CycleOffset: 0 m_Transitions: - - {fileID: 1101404391616814046} + - {fileID: -632880110147916710} + - {fileID: -6673604382440492192} + - {fileID: -70235823626180740} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -534,24 +256,25 @@ AnimatorState: m_MirrorParameterActive: 0 m_CycleOffsetParameterActive: 0 m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: e6a236f5b06f9cb44ae1ca8b50fcd8ca, type: 3} + m_Motion: {fileID: 7400000, guid: 107649059ea401b4e9c5c20f21e99a55, type: 3} m_Tag: m_SpeedParameter: m_MirrorParameter: m_CycleOffsetParameter: m_TimeParameter: ---- !u!1102 &1102379806414260514 +--- !u!1102 &-4187437994059944167 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: WalkBack_Shoot_AR - m_Speed: 1 + m_Name: Fire + m_Speed: 10 m_CycleOffset: 0 m_Transitions: - - {fileID: 1101018267375076786} + - {fileID: -146642472328627549} + - {fileID: -8614502741554326989} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -561,132 +284,128 @@ AnimatorState: m_MirrorParameterActive: 0 m_CycleOffsetParameterActive: 0 m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 669b29b2139d8de4ca5447dd50e201af, type: 3} + m_Motion: {fileID: 7400000, guid: 1d4365e1541bb6949a273318862b72d3, type: 3} m_Tag: m_SpeedParameter: m_MirrorParameter: m_CycleOffsetParameter: m_TimeParameter: ---- !u!1102 &1102525727787806748 -AnimatorState: +--- !u!1101 &-4114131529631250501 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: -5023192667791512651} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.7343743 + m_TransitionOffset: 0.08256394 + m_ExitTime: 0.28710982 + m_HasExitTime: 0 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &-3812287898245291883 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Jump to Run + m_Conditions: [] + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 8354844821256608690} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.20300466 + m_TransitionOffset: 0 + m_ExitTime: 0.7916667 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &-3071590976036615157 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 381506674367370628} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0.542241 + m_ExitTime: 0.75 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1107 &-2302487397917704150 +AnimatorStateMachine: serializedVersion: 6 m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: WalkLeft_Shoot_AR - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101624051492446352} + m_Name: Lower + m_ChildStates: + - serializedVersion: 1 + m_State: {fileID: 8354844821256608690} + m_Position: {x: 460, y: 60, z: 0} + - serializedVersion: 1 + m_State: {fileID: -5914497066343941395} + m_Position: {x: 580, y: -120, z: 0} + - serializedVersion: 1 + m_State: {fileID: -1946935121236979824} + m_Position: {x: 490, y: 230, z: 0} + - serializedVersion: 1 + m_State: {fileID: 381506674367370628} + m_Position: {x: 980, y: 60, z: 0} + m_ChildStateMachines: [] + m_AnyStateTransitions: [] + m_EntryTransitions: [] + m_StateMachineTransitions: {} m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 34d1b8e01a0412f449dcefc77b05e27b, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102593466183913558 + m_AnyStatePosition: {x: 50, y: 20, z: 0} + m_EntryPosition: {x: 50, y: 120, z: 0} + m_ExitPosition: {x: 1130, y: 240, z: 0} + m_ParentStateMachinePosition: {x: 800, y: 20, z: 0} + m_DefaultState: {fileID: 8354844821256608690} +--- !u!1102 &-1946935121236979824 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: Idle_Guard_AR + m_Name: Crouch m_Speed: 1 m_CycleOffset: 0 m_Transitions: - - {fileID: 1101567676818271368} - m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 21a805ef27ad7ef42904fff42be01ba6, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102625340767904258 -AnimatorState: - serializedVersion: 6 - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: Reload - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101363450263446180} - m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 45b9c258f262e844e8f5316722fc04c6, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102642198119822076 -AnimatorState: - serializedVersion: 6 - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: Shoot_BurstShot_AR - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101350354845918052} - m_StateMachineBehaviours: [] - m_Position: {x: 50, y: 50, z: 0} - m_IKOnFeet: 0 - m_WriteDefaultValues: 1 - m_Mirror: 0 - m_SpeedParameterActive: 0 - m_MirrorParameterActive: 0 - m_CycleOffsetParameterActive: 0 - m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 9e6c2d8b423b3db4ca189e8497d60f1e, type: 3} - m_Tag: - m_SpeedParameter: - m_MirrorParameter: - m_CycleOffsetParameter: - m_TimeParameter: ---- !u!1102 &1102885244088947746 -AnimatorState: - serializedVersion: 6 - m_ObjectHideFlags: 1 - m_CorrespondingSourceObject: {fileID: 0} - m_PrefabInstance: {fileID: 0} - m_PrefabAsset: {fileID: 0} - m_Name: Idle_Ducking_AR - m_Speed: 1 - m_CycleOffset: 0 - m_Transitions: - - {fileID: 1101376048922213750} + - {fileID: 5680347307725438578} + - {fileID: -3071590976036615157} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -699,21 +418,197 @@ AnimatorState: m_Motion: {fileID: 7400000, guid: 3ad7c5979f6586d4a9532a55492a0ebe, type: 3} m_Tag: m_SpeedParameter: - m_MirrorParameter: + m_MirrorParameter: _jump m_CycleOffsetParameter: m_TimeParameter: ---- !u!1102 &1102901032506865742 +--- !u!206 &-1862914767576164720 +BlendTree: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Blend Tree + m_Childs: + - serializedVersion: 2 + m_Motion: {fileID: 7400000, guid: 107649059ea401b4e9c5c20f21e99a55, type: 3} + m_Threshold: 0 + m_Position: {x: 0, y: 0} + m_TimeScale: 0.5 + m_CycleOffset: 0 + m_DirectBlendParameter: horizontal + m_Mirror: 0 + - serializedVersion: 2 + m_Motion: {fileID: 7400000, guid: 043a0882d93547c4da0104443de76efb, type: 3} + m_Threshold: 1 + m_Position: {x: 1, y: 0} + m_TimeScale: 1 + m_CycleOffset: 0 + m_DirectBlendParameter: horizontal + m_Mirror: 0 + m_BlendParameter: _vertical + m_BlendParameterY: _horizontal + m_MinThreshold: 0 + m_MaxThreshold: 1 + m_UseAutomaticThresholds: 1 + m_NormalizedBlendValues: 0 + m_BlendType: 3 +--- !u!1101 &-632880110147916710 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _jump + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 7739110899394721029} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.8125 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &-146642472328627549 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 2 + m_ConditionEvent: _fired + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: -4600760231423422918} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.8125 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &-70235823626180740 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: -5023192667791512651} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.982062 + m_TransitionOffset: 0.000000027939697 + m_ExitTime: 0.26345384 + m_HasExitTime: 0 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!91 &9100000 +AnimatorController: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: CharacterAnimator + serializedVersion: 5 + m_AnimatorParameters: + - m_Name: _horizontal + m_Type: 1 + m_DefaultFloat: 0 + m_DefaultInt: 0 + m_DefaultBool: 0 + m_Controller: {fileID: 9100000} + - m_Name: _vertical + m_Type: 1 + m_DefaultFloat: 1 + m_DefaultInt: 0 + m_DefaultBool: 0 + m_Controller: {fileID: 9100000} + - m_Name: _jump + m_Type: 4 + m_DefaultFloat: 0 + m_DefaultInt: 0 + m_DefaultBool: 0 + m_Controller: {fileID: 9100000} + - m_Name: _crouch + m_Type: 4 + m_DefaultFloat: 0 + m_DefaultInt: 0 + m_DefaultBool: 0 + m_Controller: {fileID: 9100000} + - m_Name: _died + m_Type: 4 + m_DefaultFloat: 0 + m_DefaultInt: 0 + m_DefaultBool: 0 + m_Controller: {fileID: 9100000} + - m_Name: _fired + m_Type: 4 + m_DefaultFloat: 0 + m_DefaultInt: 0 + m_DefaultBool: 0 + m_Controller: {fileID: 9100000} + m_AnimatorLayers: + - serializedVersion: 5 + m_Name: Lower + m_StateMachine: {fileID: -2302487397917704150} + m_Mask: {fileID: 31900000, guid: 1122aed799ca7574a8f0d2efa30e9d99, type: 2} + m_Motions: [] + m_Behaviours: [] + m_BlendingMode: 0 + m_SyncedLayerIndex: -1 + m_DefaultWeight: 0 + m_IKPass: 0 + m_SyncedLayerAffectsTiming: 0 + m_Controller: {fileID: 9100000} + - serializedVersion: 5 + m_Name: Upper + m_StateMachine: {fileID: -8265500127550764659} + m_Mask: {fileID: 31900000, guid: 368b178fc56a14549b588ee80c7cbf81, type: 2} + m_Motions: [] + m_Behaviours: [] + m_BlendingMode: 0 + m_SyncedLayerIndex: -1 + m_DefaultWeight: 1 + m_IKPass: 0 + m_SyncedLayerAffectsTiming: 0 + m_Controller: {fileID: 9100000} +--- !u!1102 &381506674367370628 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: Run_guard_AR + m_Name: Die m_Speed: 1 m_CycleOffset: 0 m_Transitions: - - {fileID: 1101995184118313822} + - {fileID: -4911913766117122026} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -723,24 +618,222 @@ AnimatorState: m_MirrorParameterActive: 0 m_CycleOffsetParameterActive: 0 m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 043a0882d93547c4da0104443de76efb, type: 3} + m_Motion: {fileID: 7400000, guid: d406f8f3cbe268f4e9d0234d45cca60c, type: 3} m_Tag: m_SpeedParameter: m_MirrorParameter: m_CycleOffsetParameter: m_TimeParameter: ---- !u!1102 &1102942199553133880 +--- !u!1101 &1131199853383832992 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 2 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 0} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 1 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.765625 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &1426403871767708545 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 2 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 0} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 1 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.765625 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &1766918516916494365 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _crouch + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: -1946935121236979824} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.25 + m_TransitionOffset: 0 + m_ExitTime: 0.75 + m_HasExitTime: 0 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &2942481661802285141 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 381506674367370628} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.45466095 + m_TransitionOffset: 0 + m_ExitTime: 0.5480226 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &4669644873837644826 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 381506674367370628} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 1.0268358 + m_TransitionOffset: 0.026483208 + m_ExitTime: 0.22033934 + m_HasExitTime: 0 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &5168308916736617153 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: Jump to Run + m_Conditions: [] + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 8354844821256608690} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.20300466 + m_TransitionOffset: 0 + m_ExitTime: 0.7916667 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &5680347307725438578 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 2 + m_ConditionEvent: _crouch + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 8354844821256608690} + m_Solo: 1 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.09322041 + m_TransitionOffset: 0 + m_ExitTime: 0.75 + m_HasExitTime: 0 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1101 &7161422939700495704 +AnimatorStateTransition: + m_ObjectHideFlags: 1 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 381506674367370628} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.45466095 + m_TransitionOffset: 0 + m_ExitTime: 0.5480226 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 +--- !u!1102 &7739110899394721029 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: Run_gunMiddle_AR + m_Name: Jump m_Speed: 1 m_CycleOffset: 0 m_Transitions: - - {fileID: 1101213140064716440} + - {fileID: -5341886129914063569} + - {fileID: -4114131529631250501} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -750,24 +843,26 @@ AnimatorState: m_MirrorParameterActive: 0 m_CycleOffsetParameterActive: 0 m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: d805ca1ced9a1054c9cb7e1dc431a6e3, type: 3} + m_Motion: {fileID: 7400000, guid: c09541f4236345c4fa4e4745793a59f3, type: 3} m_Tag: m_SpeedParameter: m_MirrorParameter: m_CycleOffsetParameter: m_TimeParameter: ---- !u!1102 &1102949299928912516 +--- !u!1102 &8354844821256608690 AnimatorState: serializedVersion: 6 m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: Shoot_SingleShot_AR + m_Name: Locomotion m_Speed: 1 m_CycleOffset: 0 m_Transitions: - - {fileID: 1101616875765159898} + - {fileID: -4865886577319040672} + - {fileID: 1766918516916494365} + - {fileID: 4669644873837644826} m_StateMachineBehaviours: [] m_Position: {x: 50, y: 50, z: 0} m_IKOnFeet: 0 @@ -777,76 +872,34 @@ AnimatorState: m_MirrorParameterActive: 0 m_CycleOffsetParameterActive: 0 m_TimeParameterActive: 0 - m_Motion: {fileID: 7400000, guid: 7e58f38d10fcc9a498d348e7339ce3ec, type: 3} + m_Motion: {fileID: -1862914767576164720} m_Tag: m_SpeedParameter: m_MirrorParameter: m_CycleOffsetParameter: m_TimeParameter: ---- !u!1107 &1107476691846109156 -AnimatorStateMachine: - serializedVersion: 6 +--- !u!1101 &9141976730198879995 +AnimatorStateTransition: m_ObjectHideFlags: 1 m_CorrespondingSourceObject: {fileID: 0} m_PrefabInstance: {fileID: 0} m_PrefabAsset: {fileID: 0} - m_Name: Base Layer - m_ChildStates: - - serializedVersion: 1 - m_State: {fileID: 1102593466183913558} - m_Position: {x: 300, y: 108, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102238716766734490} - m_Position: {x: 300, y: 168, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102100670134039114} - m_Position: {x: 300, y: 228, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102277268960662982} - m_Position: {x: 280, y: 440, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102885244088947746} - m_Position: {x: 300, y: 288, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102088155696507906} - m_Position: {x: 516, y: 444, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102625340767904258} - m_Position: {x: 528, y: 300, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102901032506865742} - m_Position: {x: 768, y: 444, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102949299928912516} - m_Position: {x: 528, y: 108, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102642198119822076} - m_Position: {x: 528, y: 168, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102106218193099826} - m_Position: {x: 528, y: 228, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102243588419055754} - m_Position: {x: 768, y: 108, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102525727787806748} - m_Position: {x: 768, y: 168, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102361151154106906} - m_Position: {x: 768, y: 228, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102379806414260514} - m_Position: {x: 768, y: 300, z: 0} - - serializedVersion: 1 - m_State: {fileID: 1102942199553133880} - m_Position: {x: 768, y: 372, z: 0} - m_ChildStateMachines: [] - m_AnyStateTransitions: [] - m_EntryTransitions: [] - m_StateMachineTransitions: {} - m_StateMachineBehaviours: [] - m_AnyStatePosition: {x: 50, y: 20, z: 0} - m_EntryPosition: {x: 50, y: 120, z: 0} - m_ExitPosition: {x: 48, y: 72, z: 0} - m_ParentStateMachinePosition: {x: 800, y: 20, z: 0} - m_DefaultState: {fileID: 1102593466183913558} + m_Name: + m_Conditions: + - m_ConditionMode: 1 + m_ConditionEvent: _died + m_EventTreshold: 0 + m_DstStateMachine: {fileID: 0} + m_DstState: {fileID: 381506674367370628} + m_Solo: 0 + m_Mute: 0 + m_IsExit: 0 + serializedVersion: 3 + m_TransitionDuration: 0.45466095 + m_TransitionOffset: 0 + m_ExitTime: 0.5480226 + m_HasExitTime: 1 + m_HasFixedDuration: 1 + m_InterruptionSource: 0 + m_OrderedInterruption: 1 + m_CanTransitionToSelf: 1 diff --git a/Assets/Defaults/mat_proBuilder.mat.meta b/Assets/Scripts/Animators/Leonid Animator/CharacterAnimator.controller.meta old mode 100755 new mode 100644 similarity index 64% rename from Assets/Defaults/mat_proBuilder.mat.meta rename to Assets/Scripts/Animators/Leonid Animator/CharacterAnimator.controller.meta index 957f067..63abf20 --- a/Assets/Defaults/mat_proBuilder.mat.meta +++ b/Assets/Scripts/Animators/Leonid Animator/CharacterAnimator.controller.meta @@ -1,8 +1,8 @@ fileFormatVersion: 2 -guid: 536ae120520e6f245bad32b94a03f6a2 +guid: 3ebf60422b6cb1c498ee4cf238072b43 NativeFormatImporter: externalObjects: {} - mainObjectFileID: 2100000 + mainObjectFileID: 9100000 userData: assetBundleName: assetBundleVariant: diff --git a/Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs b/Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs new file mode 100644 index 0000000..af2c968 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs @@ -0,0 +1,105 @@ +using UnityEngine; + +namespace Animators.Leonid_Animator +{ + [RequireComponent( + typeof(Rigidbody), + typeof(InputHandler), + typeof(AnimatorHandler))] + public class CharacterLocomotion : MonoBehaviour + { + private Transform _cameraObject; + private InputHandler _inputHandler; + private Vector3 _moveDirection; + + [HideInInspector] public Transform myTransform; + [HideInInspector] public AnimatorHandler myAnimatorHandler; + + public Rigidbody myRigidbody; + [SerializeField] public float jumpForce; + public GameObject normalCamera; + + [Header("Stats")] + [SerializeField] private float movementSpeed = 5; + [SerializeField] private float rotationSpeed = 10; + + private void Start() + { + myRigidbody = GetComponent(); + _inputHandler = GetComponent(); + myAnimatorHandler = GetComponent(); + + _cameraObject = Camera.main.transform; + myTransform = transform; + myAnimatorHandler.Initialize(); + } + + private void Update() + { + var deltaTime = Time.deltaTime; + _inputHandler.TickInput(deltaTime); + _moveDirection = _cameraObject.forward * _inputHandler.vertical + + _cameraObject.right * _inputHandler.horizontal; + _moveDirection.Normalize(); + _moveDirection *= movementSpeed; + _moveDirection.y = 0; + + var projectedVelocity = Vector3.ProjectOnPlane(_moveDirection, _normalVector); + myRigidbody.velocity = projectedVelocity; + + if (myAnimatorHandler.canRotate) + { + HandleRotation(deltaTime); + } + + myAnimatorHandler.UpdateAnimatorValues( + _inputHandler.moveAmount, + 0, + _inputHandler.jumpPressed, + _inputHandler.crouchPressed, + _inputHandler.firePressed); + + var velocity = myRigidbody.velocity; + myRigidbody.AddForce(_inputHandler.jumpPressed ? + new Vector3(0, jumpForce, 0) + : new Vector3(velocity.x*100, -50, velocity.z * 100)); + } + + private void LateUpdate() + { + _inputHandler.jumpPressed = false; + _inputHandler.crouchPressed = false; + } + + #region Movement + + private Vector3 _normalVector; + private Vector3 _targetPosition; + + private void HandleRotation(float delta) + { + if (Mathf.Abs(_inputHandler.horizontal) + Mathf.Abs(_inputHandler.vertical) < 0.1) + { + print("stop"); + return; + } + + print("begin"); + var moveAmount = _inputHandler.moveAmount; + var targetDir = _cameraObject.forward * _inputHandler.vertical + + _cameraObject.right * _inputHandler.horizontal; + targetDir.Normalize(); + targetDir.y = 0; + if (targetDir == Vector3.zero) + targetDir = myTransform.forward; + + var rotSpeed = rotationSpeed; + + var rotation = Quaternion.LookRotation(targetDir); + var targetRotation = Quaternion.Slerp(myTransform.rotation, rotation, rotationSpeed * delta); + myTransform.rotation = targetRotation; + } + + #endregion + } +} \ No newline at end of file diff --git a/Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs.meta b/Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs.meta new file mode 100644 index 0000000..715f6df --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/CharacterLocomotion.cs.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: aeafb7b8074141969e8779cd3d4a9d08 +timeCreated: 1652026088 \ No newline at end of file diff --git a/Assets/Scripts/Animators/Leonid Animator/InputHandler.cs b/Assets/Scripts/Animators/Leonid Animator/InputHandler.cs new file mode 100644 index 0000000..3e971e7 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/InputHandler.cs @@ -0,0 +1,77 @@ +using CameraScripts; +using UnityEngine; + +namespace Animators.Leonid_Animator +{ + public class InputHandler : MonoBehaviour + { + public float horizontal; + public float vertical; + public float moveAmount; + public float mouseX; + public float mouseY; + + public bool crouchPressed; + public bool jumpPressed; + public bool firePressed; + + private ThirdPersonViewInput _inputActions; + private Vector2 _movementInput; + private Vector2 _cameraInput; + + private CameraHandler _cameraHandler; + + private void Awake() + { + _cameraHandler = CameraHandler.Singleton; + if (_cameraHandler == null) + Debug.LogError("Camera Handler not found"); + } + + private void Update() + { + _cameraHandler.TargetPosition(Time.deltaTime); + _cameraHandler.HandleCameraRotation(Time.deltaTime, mouseX, mouseY); + } + + private void OnEnable() + { + if (_inputActions is null) + { + _inputActions = new ThirdPersonViewInput(); + _inputActions.PlayerMovement.Movement.performed += + context => _movementInput = context.ReadValue(); + _inputActions.PlayerMovement.Camera.performed += + context => _cameraInput = context.ReadValue(); + _inputActions.PlayerActions.Crouch.performed += + context => crouchPressed = true; + _inputActions.PlayerActions.Jump.performed += + context => jumpPressed = true; + _inputActions.PlayerActions.Fire.performed += + context => firePressed = true; + _inputActions.PlayerActions.Fire.canceled += + context => firePressed = false; + } + _inputActions.Enable(); + } + + private void OnDisable() + { + _inputActions.Disable(); + } + + public void TickInput(float delta) + { + MoveInput(delta); + } + + private void MoveInput(float delta) + { + horizontal = _movementInput.x; + vertical = _movementInput.y; + moveAmount = Mathf.Clamp01(Mathf.Abs(horizontal) + Mathf.Abs(vertical)); + mouseX = _cameraInput.x; + mouseY = _cameraInput.y; + } + } +} \ No newline at end of file diff --git a/Assets/Scripts/Animators/Leonid Animator/InputHandler.cs.meta b/Assets/Scripts/Animators/Leonid Animator/InputHandler.cs.meta new file mode 100644 index 0000000..c4d1f8e --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/InputHandler.cs.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: 77c38ddfaba349c590d4a6583f7efac4 +timeCreated: 1652025145 \ No newline at end of file diff --git a/Assets/Scripts/Animators/Leonid Animator/LowerBody.mask b/Assets/Scripts/Animators/Leonid Animator/LowerBody.mask new file mode 100644 index 0000000..ebcf1fa --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/LowerBody.mask @@ -0,0 +1,139 @@ +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!319 &31900000 +AvatarMask: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: LowerBody + m_Mask: 01000000010000000100000001000000010000000100000001000000010000000100000001000000010000000100000001000000 + m_Elements: + - m_Path: + m_Weight: 1 + - m_Path: Arm1 + m_Weight: 1 + - m_Path: AssaultRifle + m_Weight: 1 + - m_Path: Backpack1 + m_Weight: 1 + - m_Path: Body1 + m_Weight: 1 + - m_Path: head1 + m_Weight: 1 + - m_Path: Hips + m_Weight: 1 + - m_Path: Hips/ArmPosition_Left + m_Weight: 1 + - m_Path: Hips/ArmPosition_Right + m_Weight: 1 + - m_Path: Hips/ArmPosition_Right/magazine_Right + m_Weight: 1 + - m_Path: Hips/ArmPosition_Right/Trigger_Right + m_Weight: 1 + - m_Path: Hips/Spine + m_Weight: 1 + - m_Path: Hips/Spine/Chest + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack/ArmPlacement_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack/ArmPlacement_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack/ArmPlacement_Upper + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck/Head + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck/Head/Headgear_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck/Head/Headgear_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/ShoulderPadCTRL_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/ShoulderPadCTRL_Left/ShoulderPadBlade_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/ShoulderPadCTRL_Left/ShoulderPadBody_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Index_Proximal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Index_Proximal_Left/Index_Intermediate_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Index_Proximal_Left/Index_Intermediate_Left/Index_Distal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/RestOfFingers_Proximal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/RestOfFingers_Proximal_Left/RestOfFingers_Intermediate_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/RestOfFingers_Proximal_Left/RestOfFingers_Intermediate_Left/RestOfFingers_Distal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Thumb_Proximal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Thumb_Proximal_Left/Thumb_Intermediate_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Thumb_Proximal_Left/Thumb_Intermediate_Left/Thumb_Distal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/ShoulderPadCTRL_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/ShoulderPadCTRL_Right/ShoulderPadBlade_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/ShoulderPadCTRL_Right/ShoulderPadBody_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Index_Proximal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Index_Proximal_Right/Index_Intermediate_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Index_Proximal_Right/Index_Intermediate_Right/Index_Distal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/RestOfFingers_Proximal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/RestOfFingers_Proximal_Right/RestOfFingers_Intermediate_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/RestOfFingers_Proximal_Right/RestOfFingers_Intermediate_Right/RestOfFingers_Distal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Thumb_Proximal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Thumb_Proximal_Right/Thumb_Intermediate_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Thumb_Proximal_Right/Thumb_Intermediate_Right/Thumb_Distal_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left/Foot_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left/Foot_Left/Toe_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left/Foot_Left/Toe_Left/Toetip_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right/Foot_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right/Foot_Right/Toe_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right/Foot_Right/Toe_Right/Toetip_Right + m_Weight: 1 + - m_Path: Leg1 + m_Weight: 1 diff --git a/Assets/Scripts/Animators/Leonid Animator/LowerBody.mask.meta b/Assets/Scripts/Animators/Leonid Animator/LowerBody.mask.meta new file mode 100644 index 0000000..ef5e183 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/LowerBody.mask.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: 1122aed799ca7574a8f0d2efa30e9d99 +NativeFormatImporter: + externalObjects: {} + mainObjectFileID: 31900000 + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Scripts/Animators/Leonid Animator/UpperBody.mask b/Assets/Scripts/Animators/Leonid Animator/UpperBody.mask new file mode 100644 index 0000000..0384688 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/UpperBody.mask @@ -0,0 +1,139 @@ +%YAML 1.1 +%TAG !u! tag:unity3d.com,2011: +--- !u!319 &31900000 +AvatarMask: + m_ObjectHideFlags: 0 + m_CorrespondingSourceObject: {fileID: 0} + m_PrefabInstance: {fileID: 0} + m_PrefabAsset: {fileID: 0} + m_Name: UpperBody + m_Mask: 00000000010000000100000000000000000000000100000001000000010000000100000000000000000000000000000000000000 + m_Elements: + - m_Path: + m_Weight: 1 + - m_Path: Arm1 + m_Weight: 1 + - m_Path: AssaultRifle + m_Weight: 1 + - m_Path: Backpack1 + m_Weight: 1 + - m_Path: Body1 + m_Weight: 1 + - m_Path: head1 + m_Weight: 1 + - m_Path: Hips + m_Weight: 1 + - m_Path: Hips/ArmPosition_Left + m_Weight: 1 + - m_Path: Hips/ArmPosition_Right + m_Weight: 1 + - m_Path: Hips/ArmPosition_Right/magazine_Right + m_Weight: 1 + - m_Path: Hips/ArmPosition_Right/Trigger_Right + m_Weight: 1 + - m_Path: Hips/Spine + m_Weight: 1 + - m_Path: Hips/Spine/Chest + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack/ArmPlacement_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack/ArmPlacement_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/BackPack/ArmPlacement_Upper + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck/Head + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck/Head/Headgear_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Neck/Head/Headgear_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/ShoulderPadCTRL_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/ShoulderPadCTRL_Left/ShoulderPadBlade_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/ShoulderPadCTRL_Left/ShoulderPadBody_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Index_Proximal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Index_Proximal_Left/Index_Intermediate_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Index_Proximal_Left/Index_Intermediate_Left/Index_Distal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/RestOfFingers_Proximal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/RestOfFingers_Proximal_Left/RestOfFingers_Intermediate_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/RestOfFingers_Proximal_Left/RestOfFingers_Intermediate_Left/RestOfFingers_Distal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Thumb_Proximal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Thumb_Proximal_Left/Thumb_Intermediate_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Left/UpperArm_Left/LowerArm_Left/Hand_Left/Thumb_Proximal_Left/Thumb_Intermediate_Left/Thumb_Distal_Left + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/ShoulderPadCTRL_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/ShoulderPadCTRL_Right/ShoulderPadBlade_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/ShoulderPadCTRL_Right/ShoulderPadBody_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Index_Proximal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Index_Proximal_Right/Index_Intermediate_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Index_Proximal_Right/Index_Intermediate_Right/Index_Distal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/RestOfFingers_Proximal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/RestOfFingers_Proximal_Right/RestOfFingers_Intermediate_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/RestOfFingers_Proximal_Right/RestOfFingers_Intermediate_Right/RestOfFingers_Distal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Thumb_Proximal_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Thumb_Proximal_Right/Thumb_Intermediate_Right + m_Weight: 1 + - m_Path: Hips/Spine/Chest/Shoulder_Right/UpperArm_Right/LowerArm_Right/Hand_Right/Thumb_Proximal_Right/Thumb_Intermediate_Right/Thumb_Distal_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left/Foot_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left/Foot_Left/Toe_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Left/LowerLeg_Left/Foot_Left/Toe_Left/Toetip_Left + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right/Foot_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right/Foot_Right/Toe_Right + m_Weight: 1 + - m_Path: Hips/UpperLeg_Right/LowerLeg_Right/Foot_Right/Toe_Right/Toetip_Right + m_Weight: 1 + - m_Path: Leg1 + m_Weight: 1 diff --git a/Assets/Scripts/Animators/Leonid Animator/UpperBody.mask.meta b/Assets/Scripts/Animators/Leonid Animator/UpperBody.mask.meta new file mode 100644 index 0000000..42f9f75 --- /dev/null +++ b/Assets/Scripts/Animators/Leonid Animator/UpperBody.mask.meta @@ -0,0 +1,8 @@ +fileFormatVersion: 2 +guid: 368b178fc56a14549b588ee80c7cbf81 +NativeFormatImporter: + externalObjects: {} + mainObjectFileID: 31900000 + userData: + assetBundleName: + assetBundleVariant: diff --git a/Assets/Scripts/CameraScripts.meta b/Assets/Scripts/CameraScripts.meta new file mode 100644 index 0000000..28c64a8 --- /dev/null +++ b/Assets/Scripts/CameraScripts.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: 563fa8c0f982459e8a6357c9f9078744 +timeCreated: 1652086279 \ No newline at end of file diff --git a/Assets/Scripts/CameraScripts/CameraHandler.cs b/Assets/Scripts/CameraScripts/CameraHandler.cs new file mode 100644 index 0000000..d9d4672 --- /dev/null +++ b/Assets/Scripts/CameraScripts/CameraHandler.cs @@ -0,0 +1,61 @@ +using System; +using Unity.Mathematics; +using UnityEngine; + +namespace CameraScripts +{ + public class CameraHandler : MonoBehaviour + { + public Transform targetTransform; + public Transform cameraTransform; + public Transform cameraPivotTransform; + private Transform _myTransform; + private Vector3 _cameraTransformPosition; + private LayerMask ignoreLayers = ~(1 << 8 | 1 << 9 | 1 << 10); + + public static CameraHandler Singleton; + + public const float LookSpeed = 0.1f; + public const float FollowSpeed = 0.1f; + public const float PivotSpeed = 0.03f; + + private float _defaultPosition; + private float _lookAngle; + private float _pivotAngle; + + public float minimumPivot = -35; + public float maximumPivot = 35; + + + private void Awake() + { + Application.targetFrameRate = 60; + Singleton = this; + _myTransform = transform; + _defaultPosition = _myTransform.localPosition.z; + } + + public void TargetPosition(float delta) + { + var toTargetPosition = Vector3.Lerp(_myTransform.position, targetTransform.position, delta /FollowSpeed); + _myTransform.position = toTargetPosition; + } + + public void HandleCameraRotation(float delta, float mouseX, float mouseY) + { + _lookAngle += (mouseX * LookSpeed) / delta; + _pivotAngle -= (mouseY * PivotSpeed) / delta; + _pivotAngle = Mathf.Clamp(_pivotAngle, minimumPivot, maximumPivot); + + var rotation = Vector3.zero; + rotation.y = _lookAngle; + var targetRotation = Quaternion.Euler(rotation); + _myTransform.rotation = targetRotation; + + rotation = Vector3.zero; + rotation.x = _pivotAngle; + targetRotation = Quaternion.Euler(rotation); + cameraPivotTransform.localRotation = targetRotation; + } + } +} \ No newline at end of file diff --git a/Assets/Scripts/CameraScripts/CameraHandler.cs.meta b/Assets/Scripts/CameraScripts/CameraHandler.cs.meta new file mode 100644 index 0000000..b682885 --- /dev/null +++ b/Assets/Scripts/CameraScripts/CameraHandler.cs.meta @@ -0,0 +1,3 @@ +fileFormatVersion: 2 +guid: 3d606407023147d7b4d530a9593e9697 +timeCreated: 1652086288 \ No newline at end of file diff --git a/Assets/Scripts/Character/NPC.cs b/Assets/Scripts/Character/NPC.cs index 7af2ae9..9c62f45 100644 --- a/Assets/Scripts/Character/NPC.cs +++ b/Assets/Scripts/Character/NPC.cs @@ -1,5 +1,6 @@ using System; using System.Collections.Generic; +using System.Threading.Tasks; using Unity.MLAgents; using Unity.MLAgents.Actuators; using Unity.MLAgents.Sensors; diff --git a/Assets/Scripts/Character/scr_Models.cs b/Assets/Scripts/Character/scr_Models.cs deleted file mode 100755 index c735db5..0000000 --- a/Assets/Scripts/Character/scr_Models.cs +++ /dev/null @@ -1,75 +0,0 @@ -using System; -using UnityEngine; - -public static class scr_Models -{ - #region Player - - public enum PlayerStance - { - Stand, - Crouch, - Prone - } - - [Serializable] - public class PlayerSettingsModel - { - [Header("View Settings")] - public float ViewXSensetivity; - public float ViewYSensetivity; - - public bool ViewXInverted; - public bool ViewYInverted; - - [Header("Movement Settings")] - public bool SprintingHold; - public float MovementSmoothing; - - [Header("Movement - Running")] - public float RunningForwardSpeed; - public float RunningStrafeSpeed; - - [Header("Movement - Walking")] - public float WalkingForwardSpeed; - public float WalkingBackwardSpeed; - public float WalkingStrafeSpeed; - - [Header("Jumping")] - public float JumpingHeight; - public float JumpingFalloff; - public float FallingSmoothing; - - [Header("Speed Effectors")] - public float SpeedEffector = 1; - public float CrouchSpeedEffector; - public float ProneSpeedEffector; - public float FallingSpeedEffector; - } - - [Serializable] - public class CharacterStance - { - public float CameraHeight; - public CapsuleCollider StanceCollider; - } - - #endregion - - #region - Weapons - - - [Serializable] - public class WeaponSettingsModel - { - [Header("Sway")] - public float SwayAmount; - public bool SwayYInverted; - public bool SwayXInverted; - public float SwaySmoothing; - public float SwayResetSmoothing; - public float SwayClampX; - public float SwayClampY; - } - - #endregion -} diff --git a/Assets/Scripts/Misc/NavPoint.cs b/Assets/Scripts/Misc/NavPoint.cs index a70fd61..bddfadb 100755 --- a/Assets/Scripts/Misc/NavPoint.cs +++ b/Assets/Scripts/Misc/NavPoint.cs @@ -14,12 +14,11 @@ public class NavPoint : MonoBehaviour public Vector3 Position => gameObject.transform.position; public float FlagDistance { get; private set; } - public NavPointType navType = NavPointType.Direction; + [SerializeField] public NavPointType navType; - [HideInInspector] - public int PointId = 0; - public float DeathAttr = 0; - public List EnemiesSeen = new List(); + [HideInInspector] public int PointId = 0; + [HideInInspector] public float DeathAttr = 0; + [HideInInspector] public List EnemiesSeen = new List(); private void Start() { diff --git a/Assets/Scripts/Weapons/scr_WeaponController.cs b/Assets/Scripts/Weapons/WeaponController.cs old mode 100755 new mode 100644 similarity index 96% rename from Assets/Scripts/Weapons/scr_WeaponController.cs rename to Assets/Scripts/Weapons/WeaponController.cs index 613fc61..d79577a --- a/Assets/Scripts/Weapons/scr_WeaponController.cs +++ b/Assets/Scripts/Weapons/WeaponController.cs @@ -1,47 +1,46 @@ -using UnityEngine; -using static scr_Models; -public class scr_WeaponController : MonoBehaviour -{ - private scr_CharacterController characterController; - [Header("Settings")] - public WeaponSettingsModel settings; - - private bool isInitialised; - - Vector3 newWeaponRotation; - Vector3 newWeaponRotationVelocity; - - Vector3 targetWeaponRotation; - Vector3 targetWeaponRotationVelocity; - - private void Start() - { - newWeaponRotation = transform.localRotation.eulerAngles; - } - - public void Initialise(scr_CharacterController CharacterController) - { - characterController = CharacterController; - isInitialised = true; - } - - public void Update() - { - if (!isInitialised) - { - return; - } - - targetWeaponRotation.y += settings.SwayAmount * (settings.SwayXInverted ? -characterController.input_View.x : characterController.input_View.x) * Time.deltaTime; - targetWeaponRotation.x += settings.SwayAmount * (settings.SwayYInverted ? characterController.input_View.y : -characterController.input_View.y) * Time.deltaTime; - //newWeaponRotation.x = Mathf.Clamp(newWeaponRotation.x, ViewClampYMin, ViewClampYMax); - - targetWeaponRotation.x = Mathf.Clamp(targetWeaponRotation.x, -settings.SwayClampX, settings.SwayClampX); - targetWeaponRotation.y = Mathf.Clamp(targetWeaponRotation.y, -settings.SwayClampY, settings.SwayClampY); - - targetWeaponRotation = Vector3.SmoothDamp(targetWeaponRotation, Vector3.zero, ref targetWeaponRotationVelocity, settings.SwayResetSmoothing); - newWeaponRotation = Vector3.SmoothDamp(newWeaponRotation, targetWeaponRotation, ref newWeaponRotationVelocity, settings.SwaySmoothing); - - transform.localRotation = Quaternion.Euler(newWeaponRotation); - } -} +using UnityEngine; +public class scr_WeaponController : MonoBehaviour +{ + private scr_CharacterController characterController; + [Header("Settings")] + public WeaponSettingsModel settings; + + private bool isInitialised; + + Vector3 newWeaponRotation; + Vector3 newWeaponRotationVelocity; + + Vector3 targetWeaponRotation; + Vector3 targetWeaponRotationVelocity; + + private void Start() + { + newWeaponRotation = transform.localRotation.eulerAngles; + } + + public void Initialise(scr_CharacterController CharacterController) + { + characterController = CharacterController; + isInitialised = true; + } + + public void Update() + { + if (!isInitialised) + { + return; + } + + targetWeaponRotation.y += settings.SwayAmount * (settings.SwayXInverted ? -characterController.input_View.x : characterController.input_View.x) * Time.deltaTime; + targetWeaponRotation.x += settings.SwayAmount * (settings.SwayYInverted ? characterController.input_View.y : -characterController.input_View.y) * Time.deltaTime; + //newWeaponRotation.x = Mathf.Clamp(newWeaponRotation.x, ViewClampYMin, ViewClampYMax); + + targetWeaponRotation.x = Mathf.Clamp(targetWeaponRotation.x, -settings.SwayClampX, settings.SwayClampX); + targetWeaponRotation.y = Mathf.Clamp(targetWeaponRotation.y, -settings.SwayClampY, settings.SwayClampY); + + targetWeaponRotation = Vector3.SmoothDamp(targetWeaponRotation, Vector3.zero, ref targetWeaponRotationVelocity, settings.SwayResetSmoothing); + newWeaponRotation = Vector3.SmoothDamp(newWeaponRotation, targetWeaponRotation, ref newWeaponRotationVelocity, settings.SwaySmoothing); + + transform.localRotation = Quaternion.Euler(newWeaponRotation); + } +} diff --git a/Assets/Scripts/Weapons/scr_WeaponController.cs.meta b/Assets/Scripts/Weapons/WeaponController.cs.meta old mode 100755 new mode 100644 similarity index 100% rename from Assets/Scripts/Weapons/scr_WeaponController.cs.meta rename to Assets/Scripts/Weapons/WeaponController.cs.meta diff --git a/ProjectSettings/EditorBuildSettings.asset b/ProjectSettings/EditorBuildSettings.asset index 0147887..0c40c51 100755 --- a/ProjectSettings/EditorBuildSettings.asset +++ b/ProjectSettings/EditorBuildSettings.asset @@ -5,4 +5,6 @@ EditorBuildSettings: m_ObjectHideFlags: 0 serializedVersion: 2 m_Scenes: [] - m_configObjects: {} + m_configObjects: + com.unity.input.settings: {fileID: 11400000, guid: 4480bcbc35319b14588f3c0eb33e88c8, + type: 2} diff --git a/ProjectSettings/ProjectVersion.txt b/ProjectSettings/ProjectVersion.txt new file mode 100644 index 0000000..89a11d7 --- /dev/null +++ b/ProjectSettings/ProjectVersion.txt @@ -0,0 +1,2 @@ +m_EditorVersion: 2019.4.35f1 +m_EditorVersionWithRevision: 2019.4.35f1 (0462406dff2e) diff --git a/ProjectSettings/TagManager.asset b/ProjectSettings/TagManager.asset index 2730803..3a479c3 100755 --- a/ProjectSettings/TagManager.asset +++ b/ProjectSettings/TagManager.asset @@ -20,7 +20,7 @@ TagManager: - - Player - - - + - Controller - - - diff --git a/test-ml-agents.ipynb b/test-ml-agents.ipynb new file mode 100644 index 0000000..1e2d31e --- /dev/null +++ b/test-ml-agents.ipynb @@ -0,0 +1,20623 @@ +{ + "cells": [ + { + "cell_type": "code", + "execution_count": 34, + "id": "8104e2db-b1a8-40a0-a238-5d9149fd74b0", + "metadata": {}, + "outputs": [], + "source": [ + "from mlagents_envs.environment import UnityEnvironment\n", + "import mlagents_envs\n", + "import numpy as np" + ] + }, + { + "cell_type": "code", + "execution_count": 74, + "id": "6f477382-acc9-4aec-907a-7f58caf955ed", + "metadata": {}, + "outputs": [], + "source": [ + "import random" + ] + }, + { + "cell_type": "code", + "execution_count": 83, + "id": "b7f60f26-0a90-4ea5-b2c2-b5683bda56a6", + "metadata": {}, + "outputs": [], + "source": [ + "env = UnityEnvironment()" + ] + }, + { + "cell_type": "code", + "execution_count": 84, + "id": "5929b410-12c3-4bd9-b984-b2c29a76c3f3", + "metadata": {}, + "outputs": [], + "source": [ + "env.reset()" + ] + }, + { + "cell_type": "code", + "execution_count": 85, + "id": "f108ff09-9f42-4405-add3-6df941c48f8b", + "metadata": { + "scrolled": true, + "tags": [] + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 78.21462], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 80.694435], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.29597], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 90.925804], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.28178], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26389], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 83.26209], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25988], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26341], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30755], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 85.95022], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.14938], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.25333], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.273026], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.27214], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 1. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 1. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 2. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 2. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 1. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 0. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. , 0. ,\n", + " 1. , 2. , 1. , 0. , 0. , 0. , 91.2745],\n", + " dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.26514], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 1. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 2. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 1. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 1. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 0. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n", + "Step\n", + "(, )\n", + "DecisionStep(obs=[array([[0., 1., 0., 0., 0.],\n", + " [0., 1., 0., 0., 0.],\n", + " [0., 1., 1., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.],\n", + " [0., 0., 0., 0., 0.]], dtype=float32), array([ 0. , 0. , 0. , 3. , 3. , 3. ,\n", + " 0. , 2. , 2. , 1. , 0. , 0. ,\n", + " 0. , 91.30669], dtype=float32)], reward=0.0, agent_id=0, action_mask=[array([False]), array([False]), array([False])], group_id=0, group_reward=0.0)\n", + "0.0\n" + ] + }, + { + "ename": "UnityCommunicatorStoppedException", + "evalue": "Communicator has exited.", + "output_type": "error", + "traceback": [ + "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m", + "\u001b[0;31mUnityCommunicatorStoppedException\u001b[0m Traceback (most recent call last)", + "Input \u001b[0;32mIn [85]\u001b[0m, in \u001b[0;36m\u001b[0;34m()\u001b[0m\n\u001b[1;32m 1\u001b[0m \u001b[38;5;28;01mwhile\u001b[39;00m \u001b[38;5;28;01mTrue\u001b[39;00m:\n\u001b[0;32m----> 2\u001b[0m \u001b[43menv\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mstep\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 3\u001b[0m \u001b[38;5;28mprint\u001b[39m(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mStep\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 4\u001b[0m asd \u001b[38;5;241m=\u001b[39m env\u001b[38;5;241m.\u001b[39mget_steps(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124mnpc?team=0\u001b[39m\u001b[38;5;124m'\u001b[39m)\n", + "File \u001b[0;32m~/opt/miniforge3/lib/python3.9/site-packages/mlagents_envs/timers.py:305\u001b[0m, in \u001b[0;36mtimed..wrapped\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 303\u001b[0m \u001b[38;5;28;01mdef\u001b[39;00m \u001b[38;5;21mwrapped\u001b[39m(\u001b[38;5;241m*\u001b[39margs, \u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs):\n\u001b[1;32m 304\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m hierarchical_timer(func\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__qualname__\u001b[39m):\n\u001b[0;32m--> 305\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfunc\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", + "File \u001b[0;32m~/opt/miniforge3/lib/python3.9/site-packages/mlagents_envs/environment.py:350\u001b[0m, in \u001b[0;36mUnityEnvironment.step\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 348\u001b[0m outputs \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_communicator\u001b[38;5;241m.\u001b[39mexchange(step_input, \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_poll_process)\n\u001b[1;32m 349\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m outputs \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[0;32m--> 350\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m UnityCommunicatorStoppedException(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mCommunicator has exited.\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 351\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_update_behavior_specs(outputs)\n\u001b[1;32m 352\u001b[0m rl_output \u001b[38;5;241m=\u001b[39m outputs\u001b[38;5;241m.\u001b[39mrl_output\n", + "\u001b[0;31mUnityCommunicatorStoppedException\u001b[0m: Communicator has exited." + ] + } + ], + "source": [ + "while True:\n", + " env.step()\n", + " print(\"Step\")\n", + " asd = env.get_steps('npc?team=0')\n", + " print(asd)\n", + " print(asd[0][0])\n", + " _id = asd[0][0].obs[0][0][0]\n", + " print(_id)\n", + " env.set_action_for_agent('npc?team=0', 0, mlagents_envs.environment.ActionTuple(discrete=np.array([[1, 0, random.randint(0,2)]])))" + ] + }, + { + "cell_type": "code", + "execution_count": 86, + "id": "db100c84-22ab-491b-b68d-4d5c1bbc66a3", + "metadata": {}, + "outputs": [], + "source": [ + "env.close()" + ] + } + ], + "metadata": { + "kernelspec": { + "display_name": "Python 3 (ipykernel)", + "language": "python", + "name": "python3" + }, + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.9.12" + } + }, + "nbformat": 4, + "nbformat_minor": 5 +}