Future Memory Research in Soar Nate Derbinsky University of Michigan.
-
Upload
shauna-green -
Category
Documents
-
view
213 -
download
1
Transcript of Future Memory Research in Soar Nate Derbinsky University of Michigan.
![Page 1: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/1.jpg)
Future Memory Research in Soar
Nate DerbinskyUniversity of Michigan
![Page 2: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/2.jpg)
Soar Workshop 2012 - Ann Arbor, MI 2
Outline
1. Retrospective– Episodic– Semantic
2. Next Steps– Top 3– Episodic– Semantic
20 June 2012
![Page 3: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/3.jpg)
Soar Workshop 2012 - Ann Arbor, MI 3
RetrospectiveEpisodic Memory
Nuxoll (ICCM ‘04, AAAI ‘07, CSR ‘12)– Why is EpMem useful? (cognitive capabilities)– How to integrate EpMem within a cognitive architecture?
Xu (AAAI ‘10)– How can an agent make use of EpMem to develop action models?
Gorski (CSR ‘11)– To what degree can an agent learn to use EpMem?
Derbinsky, Li (ICCBR ‘09, AAMAS ‘12, AAAI ’12ab)– What are efficient & scalable algorithms for EpMem?– To what extent are they effective and efficient across a variety of tasks over long
periods of time?– To what extent do they facilitate other cognitive processes?
20 June 2012
![Page 4: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/4.jpg)
Soar Workshop 2012 - Ann Arbor, MI 4
RetrospectiveSemantic Memory
Derbinsky (ICCM ‘10, AAAI ’11)– Architectural integration [Wang; ACT-R]– How to efficiently retrieve knowledge from large
semantic stores?– How to effectively and efficiently bias ambiguous
retrievals?
Li, Derbinsky (AAAI ‘12)– To what extent does SMem facilitate other cognitive
processes?20 June 2012
![Page 5: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/5.jpg)
Soar Workshop 2012 - Ann Arbor, MI 5
Next StepsTop 3
Episodic1. Diverse usage and
evaluation2. Bounded storage – explore
interactions with retrievals, recognition, semantic learning, consolidation…
3. Dynamically representing, reasoning about, and learning from events
Semantic1. Diverse usage and
evaluation2. Automatic encoding –
explore interactions with recognition, retrievals, episodic consolidation…
3. Building up and reasoning over hierarchical semantic knowledge
20 June 2012
![Page 6: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/6.jpg)
Soar Workshop 2012 - Ann Arbor, MI 6
Next StepsEpisodic Memory: Architecture (1)
Encoding– How to provide effective episodic capabilities for
continuous modalities (e.g. visual)?
Storage– How to bound episodic-memory growth while
maintaining effective retrievals and support for other processes (e.g. recognition, semantic learning)?
20 June 2012
![Page 7: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/7.jpg)
Soar Workshop 2012 - Ann Arbor, MI 7
Next StepsEpisodic Memory: Architecture (2)
Retrieval– Beyond cardinality [and WMA], what are effective
indicators of relevance across a variety of tasks (e.g. appraisals, LTI spread)? Efficient support?
– What are effective methods to bound per-decision episodic processing (e.g. heuristic search, anytime)?
Integration– Long-term spatial-visual memory (SVS)– Consolidation to semantic memory– Support for meta-cognitive processes (e.g. recognition)
20 June 2012
![Page 8: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/8.jpg)
Soar Workshop 2012 - Ann Arbor, MI 8
Next StepsEpisodic Memory: Agents
Reasoning– How does an agent learn to utilize and weight retrievals
with other sources of [possibly conflicting] knowledge?• Meta-knowledge about [possibly flawed] retrieval processes?
– How can an agent build up, reason about, and describe “events” from primitive episodes (could be architecture)?
Efficiency– What are useful types of episodic queries across a variety of
tasks? Better task/cue analysis -> better heuristics/approximations!
20 June 2012
![Page 9: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/9.jpg)
Soar Workshop 2012 - Ann Arbor, MI 9
Next StepsSemantic Memory: Architecture (1)
Encoding– What is the source of new LTIs (e.g.
environmental/episodic regularities)?
Storage– Is it useful to forget LTIs?
20 June 2012
![Page 10: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/10.jpg)
Soar Workshop 2012 - Ann Arbor, MI 10
Next StepsSemantic Memory: Architecture (2)
Retrieval– Besides history of past use, what are effective
indicators of relevance (e.g. context)? Efficient support?
Integration– Support for meta-cognitive processes (e.g.
recognition)
20 June 2012
![Page 11: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/11.jpg)
Soar Workshop 2012 - Ann Arbor, MI 11
Next StepsSemantic Memory: Agents
Representation– What are useful representational paradigms for LTI
augmentations?– To what extent are semantic concepts “grounded?”
Reasoning– How does agent reasoning lead to hierarchical semantic
knowledge?– What are general-purpose “inference” rule sets would be
efficient and useful across a variety of tasks? Linguistic rule sets?
20 June 2012
![Page 12: Future Memory Research in Soar Nate Derbinsky University of Michigan.](https://reader036.fdocuments.in/reader036/viewer/2022083005/56649f215503460f94c39def/html5/thumbnails/12.jpg)
Soar Workshop 2012 - Ann Arbor, MI 12
Evaluation
Nuggets• EpMem and SMem have
come a long way– Useful and efficient today!
• There are many opportunities for important progress in these mechanisms
Coal• Research takes time
20 June 2012