-
Notifications
You must be signed in to change notification settings - Fork 0
/
warehouse_motap_example.py
270 lines (239 loc) · 11.1 KB
/
warehouse_motap_example.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
import warehouse
from warehouse.envs.warehouse import Warehouse
import gym
import ce
import random
import numpy as np
import random
from enum import Enum
import itertools
#
# Params
#
NUM_TASKS = 2
NUM_AGENTS = 2
# ------------------------------------------------------------------------------
# SETUP: Construct the structures for agent to recognise task progress
# ------------------------------------------------------------------------------
task_progress = {0: "initial", 1: "in_progress", 2: "success", 3: "fail"}
# Set the initial agent locations up front
# We can set the feed points up front as well because they are static
init_agent_positions = [(0, 0), (4, 0)]
size = 10
feedpoints = [(size - 1, size // 2)]
print("Feed points", feedpoints)
# ------------------------------------------------------------------------------
# Env Setup: Construct the warehouse model as an Open AI gym python environment
# ------------------------------------------------------------------------------
env: Warehouse = gym.make(
"Warehouse-v0",
initial_agent_loc=init_agent_positions,
nagents=NUM_AGENTS,
feedpoints=feedpoints,
render_mode="human",
size=size,
seed=4321,
disable_env_checker=True,
)
# We have to set the tasks racks and task feeds which depend on the number of tasks
# sample a list of ranks in the size of the number of tasks
rack_samples = random.sample([*env.warehouse_api.racks], k=NUM_TASKS * 2)
#env.warehouse_api.add_task_rack_end(0, rack_samples[0])
#env.warehouse_api.add_task_rack_start(0, rack_samples[0])
#env.warehouse_api.add_task_feed(0, feedpoints[0])
for k in range(NUM_TASKS):
env.warehouse_api.add_task_rack_end(k, rack_samples[NUM_TASKS + k])
env.warehouse_api.add_task_rack_start(k, rack_samples[0])
env.warehouse_api.add_task_feed(k, feedpoints[0])
print("env start tasks: ", env.warehouse_api.task_racks_start)
print("env start tasks: ", env.warehouse_api.task_racks_end)
print("random task racks: ", env.warehouse_api.task_racks_start)
obs = env.reset()
print("Initial observations: ", obs)
print("Agent rack positions: ", env.agent_rack_positions)
# ------------------------------------------------------------------------------
# Executor: Define a new executor which will be used to continually run agents
# ------------------------------------------------------------------------------
s_executor = ce.SerialisableExecutor(NUM_AGENTS)
# ------------------------------------------------------------------------------
# Tasks: Construct a DFA transition function and build the Mission from this
# ------------------------------------------------------------------------------
def warehouse_replenishment_task():
task = ce.DFA(list(range(0, 8)), 0, [5], [7], [6])
# attempt to goto the rack positon without carrying anything
omega = set(env.warehouse_api.words)
# The first transition determines if the label is at the rack
task.add_transition(0, "RS_NC", 1)
excluded_words = ['_'.join(x) for x in list(itertools.product(["RS", "RE", "NFR", "F"], ["P", "D", "CR", "CNR"]))]
excluded_words.append("RE_NC")
for w in excluded_words:
task.add_transition(0, f"{w}", 7)
excluded_words.append("RS_NC")
for w in omega.difference(set(excluded_words)):
task.add_transition(0, f"{w}", 0)
# The second transition determines whether the agent picked up the rack at the
# required coord
task.add_transition(1, "RS_P", 2)
excluded_words = ['_'.join(x) for x in list(itertools.product(["NFR"], ["P"]))]
for w in excluded_words:
task.add_transition(1, f"{w}", 7)
excluded_words.append("RS_P")
for w in omega.difference(set(excluded_words)):
task.add_transition(1, f"{w}", 1)
# The third transition takes the agent to the feed position while carrying
task.add_transition(2, "F_CNR", 3)
excluded_words = ['_'.join(x) for x in list(itertools.product(["F", "RS", "RE", "NFR"], ["NC", "P", "D", "CR"]))]
for w in excluded_words:
task.add_transition(2, f"{w}", 7)
excluded_words.append("F_CNR")
for w in omega.difference(set(excluded_words)):
task.add_transition(2, f"{w}", 2)
# The fourth transition takes the agent from the feed position while carrying
# back to the rack position
task.add_transition(3, "RS_CNR", 4)
excluded_words = ['_'.join(x) for x in list(itertools.product(["F", "RS", "RE", "NFR"], ["NC", "P", "D", "CR"]))]
#excluded_words.append("RS_CNR")
for w in excluded_words:
task.add_transition(3, f"{w}", 7)
excluded_words.append("RS_CNR")
for w in omega.difference(set(excluded_words)):
task.add_transition(3, f"{w}", 3)
# The fifth transition tells the agent to drop the rack at the required square
task.add_transition(4, "RS_D", 5)
for w in omega.difference(set(["RS_D"])):
task.add_transition(4, f"{w}", 4)
for w in omega:
task.add_transition(5, f"{w}", 6)
for w in omega:
task.add_transition(6, f"{w}", 6)
for w in omega:
task.add_transition(7, f"{w}", 7)
return task
def warehouse_retry_task():
task = ce.DFA(list(range(0, 8)), 0, [5], [7], [6])
# attempt to goto the rack positon without carrying anything
omega = set(env.warehouse_api.words)
# The first transition determines if the label is at the rack
task.add_transition(0, "RS_NC", 1)
excluded_words = ['_'.join(x) for x in list(itertools.product(["RS", "RE", "NFR", "F"], ["P", "D", "CR", "CNR"]))]
excluded_words.append("RE_NC")
for w in excluded_words:
task.add_transition(0, f"{w}", 7)
excluded_words.append("RS_NC")
for w in omega.difference(set(excluded_words)):
task.add_transition(0, f"{w}", 0)
# The second transition determines whether the agent picked up the rack at the
# required coord
task.add_transition(1, "RS_P", 2)
excluded_words = ['_'.join(x) for x in list(itertools.product(["NFR", "RE"], ["P"]))]
for w in excluded_words:
task.add_transition(1, f"{w}", 7)
excluded_words.append("RS_P")
for w in omega.difference(set(excluded_words)):
task.add_transition(1, f"{w}", 1)
# The third transition takes the agent to the feed position while carrying
task.add_transition(2, "F_CNR", 3)
excluded_words = ['_'.join(x) for x in list(itertools.product(["F", "RS", "RE", "NFR"], ["NC", "P", "D", "CR"]))]
for w in excluded_words:
task.add_transition(2, f"{w}", 7)
excluded_words.append("F_CNR")
for w in omega.difference(set(excluded_words)):
task.add_transition(2, f"{w}", 2)
# The fourth transition takes the agent from the feed position while carrying
# back to the rack position
task.add_transition(3, "RE_CNR", 4)
excluded_words = ['_'.join(x) for x in list(itertools.product(["F", "RS", "RE", "NFR"], ["NC", "P", "D", "CR"]))]
excluded_words.append("RS_CNR")
for w in excluded_words:
task.add_transition(3, f"{w}", 7)
excluded_words.append("RE_CNR")
for w in omega.difference(set(excluded_words)):
task.add_transition(3, f"{w}", 3)
# The fifth transition tells the agent to drop the rack at the required square
task.add_transition(4, "RE_D", 5)
for w in omega.difference(set(["RE_D"])):
task.add_transition(4, f"{w}", 4)
for w in omega:
task.add_transition(5, f"{w}", 6)
for w in omega:
task.add_transition(6, f"{w}", 6)
for w in omega:
task.add_transition(7, f"{w}", 7)
return task
# Initialise the mission
mission = ce.Mission()
# In this test there is only one task
dfa = warehouse_replenishment_task()
dfa_retry = warehouse_retry_task()
# Add the task to the mission
for k in range(NUM_TASKS):
#mission.add_task(dfa)
mission.add_task(dfa_retry)
# specify the storage outputs for the executor to access data accoss the MOTAP interface
# ------------------------------------------------------------------------------
# SCPM: Construct the SCPM structure which is a set of instructions on how to order the
# product MDPs
# ------------------------------------------------------------------------------
# Solve the product Model
scpm = ce.SCPM(mission, NUM_AGENTS, list(range(6)))
w = [0] * NUM_AGENTS + [1./ NUM_TASKS] * NUM_TASKS
#w[-1] = 1.
#w = [1./ (NUM_AGENTS + NUM_TASKS)] * (NUM_AGENTS + NUM_AGENTS)
eps = 0.0001
#target = [-80., -100., -120.] + [0.9] * 5
#target = [-80., -150.] + [0.9] * 5
target = [-80., -150.] + [0.7] * NUM_TASKS
task_map = {0: 0, 1: 1}
sol_not_found = True
while sol_not_found:
try:
tnew = ce.scheduler_synth(scpm, env.warehouse_api, w, target, eps, s_executor)
print(tnew)
sol_not_found = False
except Exception as e:
print(e)
continue
# ------------------------------------------------------------------------------
# Rendering: Render an executor
# ------------------------------------------------------------------------------
executor = s_executor.convert_to_executor(NUM_AGENTS, NUM_TASKS)
while True:
# Initialise the actions to do nothing
actions = [6] * NUM_AGENTS
for agent in range(NUM_AGENTS):
# Condition: If the agent is not working then it is available to work
# on a new task
if env.agent_task_status[agent] == env.AgentWorkingStatus.NOT_WORKING:
task = executor.get_next_task(agent)
env.agent_performing_task[agent] = task
env.states[agent] = (env.states[agent][0], 0, None)
if task is not None:
# Update the agent as working on a task and store in the env
env.agent_task_status[agent] = env.AgentWorkingStatus.WORKING
if task is not None:
print("rack: ", env.warehouse_api.task_racks_end)
else:
# Check if the agent's task has been completed
if env.agent_performing_task[agent] is not None:
status = executor.check_done(env.agent_performing_task[agent])
if task_progress[status] in ["success", "fail"]:
print(f"Task {env.agent_performing_task[agent]} -> {task_progress[status]}")
# goto the the next task
env.agent_task_status[agent] = env.AgentWorkingStatus.NOT_WORKING
env.agent_rack_positions[agent] = None
# With the current task check what the dfa state is
if env.agent_performing_task[agent] is not None:
q = executor.dfa_current_state(env.agent_performing_task[agent])
# Set the current task in the environment
env.warehouse_api.set_task_(env.agent_performing_task[agent])
# Get the action from the scheduler stored in the executor
actions[agent] = executor.get_action(agent, env.agent_performing_task[agent], env.states[agent], q)
# step the agent forward one timestep
obs, rewards, dones, info = env.step(actions)
# Step the DFA forward
for agent in range(NUM_AGENTS):
current_task = env.agent_performing_task[agent]
if current_task is not None:
q = executor.dfa_current_state(current_task)
executor.dfa_next_state(current_task, q, info[agent]["word"])
qprime = executor.dfa_current_state(current_task)