Workshop on Self-Organization and Unsupervised Learning in Vision

Jonathan Marshall marshall at marshall.cs.unc.edu
Fri Nov 29 20:24:48 EST 1991


			    (Please post)


	    PROGRAM:  NIPS*91 Post-Conference Workshop on
				   
	SELF-ORGANIZATION AND UNSUPERVISED LEARNING IN VISION
				   
		 December 6-7, 1991 in Vail, Colorado
				   
		 Workshop Chair: Jonathan A. Marshall
				   
       Department of Computer Science, CB 3175, Sitterson Hall
   University of North Carolina, Chapel Hill, NC 27599-3175, U.S.A.   
		  919-962-1887, marshall at cs.unc.edu

Substantial neurophysiological and psychophysical evidence suggests
that visual experience guides or directs the formation of much of the
fine structure of animal visual systems.  Simple unsupervised learning
procedures (e.g., Hebbian rules) using winner-take-all or local
k-winner networks have been applied with moderate success to show how
visual experience can guide the self-organization of visual mechanisms
sensitive to low-level attributes like orientation, contrast, color,
stereo disparity, and motion.  However, such simple networks lack the
more sophisticated capabilities needed to demonstrate self-organized
development of higher-level visual mechanisms for segmentation,
grouping/binding, selective attention, representation of occluded or
amodal visual features, resolution of uncertainty, generalization,
context-sensitivity, and invariant object recognition.

A variety of enhancements to the simple Hebbian model have been
proposed.  These include anti-Hebbian rules, maximization of mutual
information, oscillatory interactions, intraneuronal interactions,
steerable receptive fields, pre- vs. post-synaptic learning rules,
covariance rules, addition of behavioral (motor) information, and
attentional gating.  Are these extensions to unsupervised learning
sufficiently powerful to model the important aspects of
neurophysiological development of higher-level visual functions?

Some of the specific questions that the workshop will address are:

  o Does our visual environment provide enough information to direct
    the formation of higher-level visual processing mechanisms?

  o What kinds of information (e.g., correlations, constraints,
    coherence, and affordances) can be discovered in our visual world,
    using unsupervised learning?

  o Can such higher-level visual processing mechanisms be formed by
    unsupervised learning?  Or is it necessary to appeal to external
    mechanisms such as evolution (genetic algorithms)?

  o Are there further enhancements that can be made to improve the
    performance and capabilities of unsupervised learning rules for
    vision?

  o What neurophysiological evidence is available regarding these
    possible enhancements to models of unsupervised learning?

  o What aspects of the development of visual systems must be
    genetically pre-wired, and what aspects can be guided or directed
    by visual experience?

  o How is the output of an unsupervised network stage used in
    subsequent stages of processing?

  o How can behaviorally relevant (sensorimotor) criteria become
    incorporated into visual processing mechanisms, using unsupervised
    learning?

This 2-day informal workshop brings together researchers in visual
neuroscience, visual psychophysics, and neural network modeling.
Invited speakers from these communities will briefly discuss their
views and results on relevant topics.  In discussion periods, we will
examine and compare these results in detail.

The workshop topic is crucial to our understanding of how animal
visual systems got the way they are.  By addressing this issue
head-on, we may come to understand better the factors that shape the
structure of animal visual systems, and we may become able to build
better computational models of the neurophysiological processes
underlying vision.


----------------------------------------------------------------------


			     PROGRAM


	   FRIDAY MORNING, December 6, 7:30-9:30 a.m.

Daniel Kersten, Department of Psychology, University of Minnesota.
  "Environmental structure and scene perception: Perceptual
   representation of material, shape, and lighting"

David C. Knill, Center for Research in Learning, Perception, and
Cognition, University of Minnesota.
  "Environmental structure and scene perception: The nature of
   visual cues for 3-D scene structure"

DISCUSSION

Edward M. Callaway, Department of Neurobiology, Duke University.
  "Development of clustered intrinsic connections in cat striate
   cortex"

Michael P. Stryker, Department of Physiology, University of
California at San Francisco.
  "Problems and promise of relating theory to experiment in models
   for the development of visual cortex"

DISCUSSION


	  FRIDAY AFTERNOON, December 6, 4:30-6:30 p.m.

Joachim M. Buhmann, Lawrence Livermore National Laboratory.
  "Complexity optimized data clustering by competitive neural
   networks"

Nicol G. Schraudolph, Department of Computer Science,
University of California at San Diego.
  "The information transparency of sigmoidal nodes"

DISCUSSION

Heinrich H. Bulthoff, Department of Cognitive and Linguistic
Sciences, Brown University.
  "Psychophysical support for a 2D view interpolation theory of
   object recognition"

John E. Hummel, Department of Psychology, University of
California at Los Angeles.
  "Structural description and self organizing object
   classification"

DISCUSSION


	  SATURDAY MORNING, December 7, 7:30-9:30 a.m.

Allan Dobbins, Computer Vision and Robotics Laboratory, McGill
University.
  "Local estimation of binocular optic flow"

Alice O'Toole, School of Human Development, The University of
Texas at Dallas.
  "Recent psychophysics suggesting a reformulation of the
   computational problem of structure-from-stereopsis"

DISCUSSION

Jonathan A. Marshall, Department of Computer Science, University
of North Carolina at Chapel Hill.
  "Development of perceptual context-sensitivity in unsupervised
   neural networks: Parsing, grouping, and segmentation"

Suzanna Becker, Department of Computer Science, University of
Toronto.
  "Learning perceptual invariants in unsupervised connectionist
   networks"

Albert L. Nigrin, Department of Computer Science and Information
Systems, American University.
  "Using Presynaptic Inhibition to Allow Neural Networks to Perform   
   Translational Invariant Recognition

DISCUSSION


	 SATURDAY AFTERNOON, December 7, 4:30-7:00 p.m.

Jurgen Schmidhuber, Department of Computer Science, University
of Colorado.
  "Learning non-redundant codes by predictability minimization"

Laurence T. Maloney, Center for Neural Science, New York
University.
  "Geometric calibration of a simple visual system"

DISCUSSION

Paul Munro, Department of Information Science, University of
Pittsburgh.
  "Self-supervised learning of concepts"

Richard Zemel, Department of Computer Science, University of
Toronto.
  "Learning to encode parts of objects"

DISCUSSION


WRAP-UP, 6:30-7:00


----------------------------------------------------------------------

			    (Please post)



More information about the Neur-sci mailing list