@@ -113,9 +113,11 @@ def new_program(percept):
113
113
action = old_program (percept )
114
114
print ('{} perceives {} and does {}' .format (agent , percept , action ))
115
115
return action
116
+
116
117
agent .program = new_program
117
118
return agent
118
119
120
+
119
121
# ______________________________________________________________________________
120
122
121
123
@@ -130,6 +132,7 @@ def program(percept):
130
132
percepts .append (percept )
131
133
action = table .get (tuple (percepts ))
132
134
return action
135
+
133
136
return program
134
137
135
138
@@ -146,26 +149,31 @@ def RandomAgentProgram(actions):
146
149
"""
147
150
return lambda percept : random .choice (actions )
148
151
152
+
149
153
# ______________________________________________________________________________
150
154
151
155
152
156
def SimpleReflexAgentProgram (rules , interpret_input ):
153
157
"""This agent takes action based solely on the percept. [Figure 2.10]"""
158
+
154
159
def program (percept ):
155
160
state = interpret_input (percept )
156
161
rule = rule_match (state , rules )
157
162
action = rule .action
158
163
return action
164
+
159
165
return program
160
166
161
167
162
168
def ModelBasedReflexAgentProgram (rules , update_state , trainsition_model , sensor_model ):
163
169
"""This agent takes action based on the percept and state. [Figure 2.12]"""
170
+
164
171
def program (percept ):
165
172
program .state = update_state (program .state , program .action , percept , trainsition_model , sensor_model )
166
173
rule = rule_match (program .state , rules )
167
174
action = rule .action
168
175
return action
176
+
169
177
program .state = program .action = None
170
178
return program
171
179
@@ -176,6 +184,7 @@ def rule_match(state, rules):
176
184
if rule .matches (state ):
177
185
return rule
178
186
187
+
179
188
# ______________________________________________________________________________
180
189
181
190
@@ -219,6 +228,7 @@ def ReflexVacuumAgent():
219
228
>>> environment.status == {(1,0):'Clean' , (0,0) : 'Clean'}
220
229
True
221
230
"""
231
+
222
232
def program (percept ):
223
233
location , status = percept
224
234
if status == 'Dirty' :
@@ -227,6 +237,7 @@ def program(percept):
227
237
return 'Right'
228
238
elif location == loc_B :
229
239
return 'Left'
240
+
230
241
return Agent (program )
231
242
232
243
@@ -253,8 +264,10 @@ def program(percept):
253
264
return 'Right'
254
265
elif location == loc_B :
255
266
return 'Left'
267
+
256
268
return Agent (program )
257
269
270
+
258
271
# ______________________________________________________________________________
259
272
260
273
@@ -392,22 +405,22 @@ def __add__(self, heading):
392
405
True
393
406
"""
394
407
if self .direction == self .R :
395
- return {
408
+ return {
396
409
self .R : Direction (self .D ),
397
410
self .L : Direction (self .U ),
398
411
}.get (heading , None )
399
412
elif self .direction == self .L :
400
- return {
413
+ return {
401
414
self .R : Direction (self .U ),
402
415
self .L : Direction (self .D ),
403
416
}.get (heading , None )
404
417
elif self .direction == self .U :
405
- return {
418
+ return {
406
419
self .R : Direction (self .R ),
407
420
self .L : Direction (self .L ),
408
421
}.get (heading , None )
409
422
elif self .direction == self .D :
410
- return {
423
+ return {
411
424
self .R : Direction (self .L ),
412
425
self .L : Direction (self .R ),
413
426
}.get (heading , None )
@@ -425,13 +438,13 @@ def move_forward(self, from_location):
425
438
"""
426
439
x , y = from_location
427
440
if self .direction == self .R :
428
- return ( x + 1 , y )
441
+ return x + 1 , y
429
442
elif self .direction == self .L :
430
- return ( x - 1 , y )
443
+ return x - 1 , y
431
444
elif self .direction == self .U :
432
- return ( x , y - 1 )
445
+ return x , y - 1
433
446
elif self .direction == self .D :
434
- return ( x , y + 1 )
447
+ return x , y + 1
435
448
436
449
437
450
class XYEnvironment (Environment ):
@@ -462,7 +475,7 @@ def things_near(self, location, radius=None):
462
475
radius2 = radius * radius
463
476
return [(thing , radius2 - distance_squared (location , thing .location ))
464
477
for thing in self .things if distance_squared (
465
- location , thing .location ) <= radius2 ]
478
+ location , thing .location ) <= radius2 ]
466
479
467
480
def percept (self , agent ):
468
481
"""By default, agent perceives things within a default radius."""
@@ -476,17 +489,17 @@ def execute_action(self, agent, action):
476
489
agent .direction += Direction .L
477
490
elif action == 'Forward' :
478
491
agent .bump = self .move_to (agent , agent .direction .move_forward (agent .location ))
479
- # elif action == 'Grab':
480
- # things = [thing for thing in self.list_things_at(agent.location)
481
- # if agent.can_grab(thing)]
482
- # if things:
483
- # agent.holding.append(things[0])
492
+ # elif action == 'Grab':
493
+ # things = [thing for thing in self.list_things_at(agent.location)
494
+ # if agent.can_grab(thing)]
495
+ # if things:
496
+ # agent.holding.append(things[0])
484
497
elif action == 'Release' :
485
498
if agent .holding :
486
499
agent .holding .pop ()
487
500
488
501
def default_location (self , thing ):
489
- return ( random .choice (self .width ), random .choice (self .height ) )
502
+ return random .choice (self .width ), random .choice (self .height )
490
503
491
504
def move_to (self , thing , destination ):
492
505
"""Move a thing to a new location. Returns True on success or False if there is an Obstacle.
@@ -505,7 +518,7 @@ def move_to(self, thing, destination):
505
518
def add_thing (self , thing , location = (1 , 1 ), exclude_duplicate_class_items = False ):
506
519
"""Add things to the world. If (exclude_duplicate_class_items) then the item won't be
507
520
added if the location has at least one item of the same class."""
508
- if ( self .is_inbounds (location ) ):
521
+ if self .is_inbounds (location ):
509
522
if (exclude_duplicate_class_items and
510
523
any (isinstance (t , thing .__class__ ) for t in self .list_things_at (location ))):
511
524
return
@@ -514,14 +527,14 @@ def add_thing(self, thing, location=(1, 1), exclude_duplicate_class_items=False)
514
527
def is_inbounds (self , location ):
515
528
"""Checks to make sure that the location is inbounds (within walls if we have walls)"""
516
529
x , y = location
517
- return not (x < self .x_start or x >= self .x_end or y < self .y_start or y >= self .y_end )
530
+ return not (x < self .x_start or x > self .x_end or y < self .y_start or y > self .y_end )
518
531
519
532
def random_location_inbounds (self , exclude = None ):
520
533
"""Returns a random location that is inbounds (within walls if we have walls)"""
521
534
location = (random .randint (self .x_start , self .x_end ),
522
535
random .randint (self .y_start , self .y_end ))
523
536
if exclude is not None :
524
- while ( location == exclude ) :
537
+ while location == exclude :
525
538
location = (random .randint (self .x_start , self .x_end ),
526
539
random .randint (self .y_start , self .y_end ))
527
540
return location
@@ -543,7 +556,7 @@ def add_walls(self):
543
556
for x in range (self .width ):
544
557
self .add_thing (Wall (), (x , 0 ))
545
558
self .add_thing (Wall (), (x , self .height - 1 ))
546
- for y in range (1 , self .height - 1 ):
559
+ for y in range (1 , self .height - 1 ):
547
560
self .add_thing (Wall (), (0 , y ))
548
561
self .add_thing (Wall (), (self .width - 1 , y ))
549
562
@@ -574,6 +587,7 @@ class Obstacle(Thing):
574
587
class Wall (Obstacle ):
575
588
pass
576
589
590
+
577
591
# ______________________________________________________________________________
578
592
579
593
@@ -682,6 +696,7 @@ def __init__(self, coordinates):
682
696
super ().__init__ ()
683
697
self .coordinates = coordinates
684
698
699
+
685
700
# ______________________________________________________________________________
686
701
# Vacuum environment
687
702
@@ -691,7 +706,6 @@ class Dirt(Thing):
691
706
692
707
693
708
class VacuumEnvironment (XYEnvironment ):
694
-
695
709
"""The environment of [Ex. 2.12]. Agent perceives dirty or clean,
696
710
and bump (into obstacle) or not; 2D discrete world of unknown size;
697
711
performance measure is 100 for each dirt cleaned, and -1 for
@@ -710,7 +724,7 @@ def percept(self, agent):
710
724
Unlike the TrivialVacuumEnvironment, location is NOT perceived."""
711
725
status = ('Dirty' if self .some_things_at (
712
726
agent .location , Dirt ) else 'Clean' )
713
- bump = ('Bump' if agent .bump else 'None' )
727
+ bump = ('Bump' if agent .bump else 'None' )
714
728
return (status , bump )
715
729
716
730
def execute_action (self , agent , action ):
@@ -729,7 +743,6 @@ def execute_action(self, agent, action):
729
743
730
744
731
745
class TrivialVacuumEnvironment (Environment ):
732
-
733
746
"""This environment has two locations, A and B. Each can be Dirty
734
747
or Clean. The agent perceives its location and the location's
735
748
status. This serves as an example of how to implement a simple
@@ -766,6 +779,7 @@ def default_location(self, thing):
766
779
"""Agents start in either location at random."""
767
780
return random .choice ([loc_A , loc_B ])
768
781
782
+
769
783
# ______________________________________________________________________________
770
784
# The Wumpus World
771
785
@@ -775,6 +789,7 @@ class Gold(Thing):
775
789
def __eq__ (self , rhs ):
776
790
"""All Gold are equal"""
777
791
return rhs .__class__ == Gold
792
+
778
793
pass
779
794
780
795
@@ -824,6 +839,7 @@ def can_grab(self, thing):
824
839
825
840
class WumpusEnvironment (XYEnvironment ):
826
841
pit_probability = 0.2 # Probability to spawn a pit in a location. (From Chapter 7.2)
842
+
827
843
# Room should be 4x4 grid of rooms. The extra 2 for walls
828
844
829
845
def __init__ (self , agent_program , width = 6 , height = 6 ):
@@ -901,12 +917,9 @@ def percept(self, agent):
901
917
"""Return things in adjacent (not diagonal) cells of the agent.
902
918
Result format: [Left, Right, Up, Down, Center / Current location]"""
903
919
x , y = agent .location
904
- result = []
905
- result .append (self .percepts_from (agent , (x - 1 , y )))
906
- result .append (self .percepts_from (agent , (x + 1 , y )))
907
- result .append (self .percepts_from (agent , (x , y - 1 )))
908
- result .append (self .percepts_from (agent , (x , y + 1 )))
909
- result .append (self .percepts_from (agent , (x , y )))
920
+ result = [self .percepts_from (agent , (x - 1 , y )), self .percepts_from (agent , (x + 1 , y )),
921
+ self .percepts_from (agent , (x , y - 1 )), self .percepts_from (agent , (x , y + 1 )),
922
+ self .percepts_from (agent , (x , y ))]
910
923
911
924
"""The wumpus gives out a loud scream once it's killed."""
912
925
wumpus = [thing for thing in self .things if isinstance (thing , Wumpus )]
@@ -949,7 +962,7 @@ def execute_action(self, agent, action):
949
962
"""The arrow travels straight down the path the agent is facing"""
950
963
if agent .has_arrow :
951
964
arrow_travel = agent .direction .move_forward (agent .location )
952
- while ( self .is_inbounds (arrow_travel ) ):
965
+ while self .is_inbounds (arrow_travel ):
953
966
wumpus = [thing for thing in self .list_things_at (arrow_travel )
954
967
if isinstance (thing , Wumpus )]
955
968
if len (wumpus ):
@@ -979,12 +992,13 @@ def is_done(self):
979
992
print ("Death by {} [-1000]." .format (explorer [0 ].killed_by ))
980
993
else :
981
994
print ("Explorer climbed out {}."
982
- .format (
983
- "with Gold [+1000]!" if Gold () not in self .things else "without Gold [+0]" ))
995
+ .format (
996
+ "with Gold [+1000]!" if Gold () not in self .things else "without Gold [+0]" ))
984
997
return True
985
998
986
-
987
999
# TODO: Arrow needs to be implemented
1000
+
1001
+
988
1002
# ______________________________________________________________________________
989
1003
990
1004
@@ -1016,13 +1030,16 @@ def test_agent(AgentFactory, steps, envs):
1016
1030
>>> result == 5
1017
1031
True
1018
1032
"""
1033
+
1019
1034
def score (env ):
1020
1035
agent = AgentFactory ()
1021
1036
env .add_thing (agent )
1022
1037
env .run (steps )
1023
1038
return agent .performance
1039
+
1024
1040
return mean (map (score , envs ))
1025
1041
1042
+
1026
1043
# _________________________________________________________________________
1027
1044
1028
1045
0 commit comments