{"title": "JPMAX: Learning to Recognize Moving Objects as a Model-fitting Problem", "book": "Advances in Neural Information Processing Systems", "page_first": 933, "page_last": 940, "abstract": null, "full_text": "JPMAX:  Learning to Recognize Moving \n\nObjects as  a  Model-fitting Problem \n\nSuzanna  Becker \n\nDepartment of Psychology,  McMaster University \n\nHamilton,  Onto  L8S  4K1 \n\nAbstract \n\nUnsupervised learning procedures have been successful at low-level \nfeature  extraction  and  preprocessing of raw  sensor  data.  So  far, \nhowever,  they  have  had  limited  success  in  learning  higher-order \nrepresentations, e.g.,  of objects in  visual images.  A promising ap(cid:173)\nproach  is  to  maximize  some  measure  of  agreement  between  the \noutputs of two groups of units which receive inputs physically sep(cid:173)\narated in  space, time or modality,  as  in  (Becker and Hinton,  1992; \nBecker, 1993; de Sa,  1993).  Using the same approach, a much sim(cid:173)\npler  learning  procedure is  proposed  here  which  discovers  features \nin a single-layer network consisting of several populations of units, \nand  can  be  applied  to  multi-layer  networks  trained  one  layer  at \na  time.  When  trained with  this  algorithm  on  image  sequences  of \nmoving geometric objects a two-layer network can learn to perform \naccurate position-invariant object classification. \n\n1  LEARNING  COHERENT CLASSIFICATIONS \n\nA powerful constraint in  sensory data is  coherence over time,  in  space,  and  across \ndifferent sensory modalities.  An  unsupervised learning procedure which can capital(cid:173)\nize on these constraints may be able to explain much of perceptual self-organization \nin  the mammalian brain.  The problem is to derive an appropriate cost function for \nunsupervised  learning which  will  capture coherence constraints in  sensory  signals; \nwe  would  also  like  it  to  be  applicable  to  multi-layer  nets  to  train  hidden  as  well \nas  output  layers.  Our  ultimate  goal  is  for  the  network  to  discover  natural object \nclasses  based on these coherence assumptions. \n\n\f934 \n\nSuzanna  Becker \n\n1.1  PREVIOUS WORK \n\nSuccessive images in  continuous visual input  are usually views  of the same object; \nthus,  although  the  image  pixels  may  change  considerably  from  frame  to  frame, \nthe image usually can  be described  by a  small set of consistent object descriptors, \nor lower-level  feature  descriptors.  We  refer  to  this  type of continuity  as  temporal \ncoherence.  This  sort  of structure is  ubiquitous  in  sensory  signals,  from  vision  as \nwell  as  other  senses,  and  can  be  used  by  a  neural  network  to  derive  temporally \ncoherent classifications.  This idea has been  used, for example, in  temporal versions \nof the Hebbian learning rule to associate items over time (Weinshall, Edelman and \nB iilt hoff,  1990; FOldiak, 1991).  To capitalize on temporal coherence for higher-order \nfeature extraction and classification,  we  need a  more powerful  learning principle. \nA promising approach is to maximize some measure of agreement between the out(cid:173)\nputs of two groups of units which receive inputs physically separated in space, time \nor modality, as in (Becker and Hinton, 1992; Becker, 1993; de Sa, 1993).  This forces \nthe units to extract features  which  are coherent across the different  input sources. \nBecker and Hinton's  (1992)  Imax algorithm maximizes  the mutual information be(cid:173)\ntween  the outputs of two  modules,  y~ and Yb,  connected  to different  parts  of the \ninput, a and b.  Becker (1993) extended this idea to the problem of classifying tem(cid:173)\nporally  varying  patterns  by  applying  the  discrete  case  of the  mutual  information \ncost  function  to the outputs of a  single module at  successive time steps,  y-;'(t)  and \ny-;'(t + 1).  However, the success of this method relied upon the back-propagation of \nderivatives  to train the  hidden  layer and it was  found  to be extremely susceptible \nto local optima.  de  Sa's method  (1993)  is  closely  related,  and minimizes  the prob(cid:173)\nability of disagreement between output classifications, y~(t) and y1(t), produced by \ntwo  modules  having  different  inputs,  e.g.,  from  different  sensory  modalities.  The \nsuccess of this method hinges  upon  bootstrapping the first  layer by initializing the \nweights to randomly selected training patterns, so this method too is susceptible to \nthe problem of local optima.  If we  had a  more flexible  cost function  that could  be \napplied  to a  multi-layer network,  first  to each  hidden  layer in  turn,  and finally  to \nthe~utput layer for  classification,  so  that the  two  layers could  discover  genuinely \ndifferent  structure,  we  might  be  able  to overcome  the  problem  of getting trapped \nin  local optima,  yielding a  more powerful  and efficient  learning procedure. \nWe  can analyze the optimal solutions for  both  de  Sa's and Becker's cost  functions \n(see Figure  1 a)  and see that both cost functions  are maximized  by having perfect \none-to-one agreement between the two groups of units over all cases, using a one-of-n \nencoding, i.e.,  having only a single output unit on for each case.  A major limitation \nof these  methods  is  that  they strive for  perfect  classifications  by  the units.  While \nthis is  desirable at the top layer of a  network,  it is  an  unsuitable  goal for  training \nintermediate layers to detect  low-level features.  For example, features like oriented \nedges would  not  be  perfect  predictors across spatially or temporally nearby image \npatches in images of translating and rotating objects.  Instead, we might expect that \nan oriented edge at one location would  predict a small range of similar orientations \nat nearby locations.  So we would prefer a cost function  whose optimal solution was \nmore like  those shown in Figure 1 b) or c).  This would allow a feature i  in  group a \nto agree with any of several nearby features,  e.g.  i  - 1,  i, or i + 1 in group  b. \n\n\fJPMAX \n\n935 \n\na) \n\n\u2022 \n\u2022 \u2022 \n=== \n.!iii \n.. \n=== ;;; \n\n......  , \n\nII \n, \n\n\u2022  == -= = \n\u2022 \n\n\u2022 \u2022  , \u2022 \u2022 \u2022 \n\nII' ill!! \nII \nI' \nI' \n\nb) \n\u2022\u2022 11 \n\u2022\u2022\u2022\u2022 \n= -.1. \nI!!!!! \u2022\u2022 \u2022 \n\u2022 \u2022 11, \u2022 \u2022 \u2022  \n\"'-= .i. \u00b7 1' \u2022\u2022\u2022 \n\u2022\u2022 \u2022 \u2022\u2022 \u2022 \n\u2022 J[., \u2022 \n\u2022 \n\u2022\u2022 \n\nii \u2022\u2022\u2022  II \n\nI \n\nc) \n\nn \n\n\u2022 \u2022  11 \n= \u2022 \u2022 \u2022 \u2022 \n= I!!!!! \u2022 I!!IJ!!' \n\u2022\u2022\u2022 \n\nII' \u2022 \u2022  \n\n:== \n\n;;; \n\nII' \n\n.. \n.. \n\u2022 \u2022 II .'. ;;; \n\u2022\u2022 .. \n== .i. Ilii \n\u2022 \n\nI \u2022  \u2022 II \n. \n\nFigure  1:  Three  possible  joint  distributions  for  the  probability  that  the i th  and j th \nunits in  two  sets  of m  classification  units are  both  on.  White  is high  density,  black \nis  low  density.  The  optimal joint  distribution  for  Becker's  and  de  Sa's  algorithms \nis  a  matrix with  all its  density  either in  the  diagonal  as  in a),  or any subset of the \ndiagonal  entries  for  de  Sa's  method,  or  a  permutation  of the  diagonal  matrix for \nBecker's  algorithm.  Alternative distributions  are  shown  in b)  and  c). \n\n1.2  THE JPMAX ALGORITHM \n\nOne way to achieve an arbitrary configuration of agreement over time between two \ngroups  of units  (as  in  Figure  1 b)  or c\u00bb \nis  to treat the desired configuration as  a \nprior  joint  probability  distribution  over  their outputs.  We  can  obtain  the  actual \ndistribution by observing the temporal correlations between  pairs of units' outputs \nin  the two  groups  over an ensemble of patterns.  We  can  then optimize the actual \ndistribution  to  fit  the  prior.  We  now  derive  two  different  cost  functions  which \nachieve this result.  Interestingly,  they result in  very similar learning rules. \nSuppose we have two groups of m units as shown in Figure 2 a), receiving inputs, x\"7t \nand xi\"  from  the same or nearby parts of the input image.  Let Ca(t)  and Cb(t)  be \nthe classifications of the two input patches produced by the network at time step t; \nthe outputs of the two groups of units, y\"7t(t)  and yi,(t),  represent these classification \nprobabilities: \n\nYai(t) \n\n-\n\nP(Ca(t) = i) = Lj eneto;(t) \n\nenetoi(t) \n\nYbi(t)  =  P(Cb(t) = i) = Lj enetb;(t) \n\nenetbi (t) \n\n(1) \n\n(the usual  \"soft max\"  output function)  where netai(t) and netbj(t) are the weighted \nnet  inputs to units.  We  could  now  observe the expected joint probability distribu(cid:173)\ntion  qij  =  E  [Yai(t)Ybj(t + 1)]t  =  E  (P(Ca(t)  =  i, Cb(t + 1) =  j)]t  by  computing the \ntemporal covariances between the classification probabilities, averaged over the en(cid:173)\nsemble of training patterns; this joint probability is  an m2-valued random variable. \nGiven  the  above  statistics,  one  possible  cost  function  we  could  minimize  is  the \n-log probability of the observed temporal covariance between the two sets of units' \noutputs  under  some  prior  distribution  (e.g.  Figure  1  b)  or  c\u00bb.  If we  knew  the \nactual frequency  counts for each (joint) classification k = kll ,\u00b7\u00b7 ., kIm, k21 ,  ... ,kmm, \n\n\f936 \n\nSuzanna Becker \n\nb) \n\nAt \n\nb'~-#--#----------\"'-\"\"\"\"-\"''''''''''''''''''''':iIo. \n\n<t (I) \nt \n\nFigure 2:  a)  Two  groups  of 15 units  receive  inputs from  a  2D  retina.  The  groups \nare  able  to  observe  each  other's  outputs  across  lateral  links  with  unit time  delays. \nb)  A  second  layer  of two  groups  of 3  units is  added  to  the  architecture  in a). \n\nrather than just the observed joint probabilities,  qij =  E  [~] ,  then given our prior \nmodel, pu, ... ,Pmm, we  could compute the probability of the observations under a \nmultinomial  distribution: \n\nUsing the de  Moivre-Laplace approximation leads  to the following: \n\nP(k) ~ \n\n1 \n\nv(27rn)m 2 -1 IL,j Pij \n\nexp (_! L (kij - n Pij )2) \n\n2  i,j \n\nnpij \n\n(2) \n\n(3) \n\nTaking the derivative of the - log probability wrt kij leads to a very simple learning \nrule which  depends only on the observed probabilities  qij  and priors Pij: \n\nfJ  -logP(k) \n\nfJkij \n\nnpij -\n\nk ij \n\nnpij \n\nPij - qij \n\n-\n\n~ \n\nPij \n\n(4) \n\nTo obtain the final weight update rule, we just multiply this by n %!i:l\n.  One problem \nwith the above formulation  is  that the priors Pij  must  not  be too close  to zero for \nthe de Moivre-Laplace approximation to hold.  In practice, this cost function  works \nwell if we  simply ignore the derivative terms  where the priors are zero. \nAn alternative cost function (as suggested by Peter Dayan) which works equally well \nis the Kullback-Liebler divergence or G-error between the desired joint probabilities \nPij  and the observed probabilities  qij: \n\nG(p,q) =  - LL (Pij logpij - Pij lOgqij) \n\nj \n\n(5) \n\n\fJPMAX \n\n937 \n\nFigure 3:  10 of the  1500 tmining patterns:  geometric  objects  centered in 36 possible \nlocations on a 12-by-12 pixel grid.  Object location varied mndomly between patterns. \n\nThe derivative of G  wrt qij  is: \n\naG \naqij \n\nPij \n\nqij \n\n(6) \n\nsubject  to  Llij qij  =  1  (enforced  by  the softmax output function). \nlarity between the learning rules given by equations 4 and 6. \n\nNote  the simi-\n\n2  EXPERIMENTS \n\nThe  network  shown  in  Figure  2 a)  was  trained  to  minimize  equation  5 on  an en(cid:173)\nsemble of pattern trajectories of circles, squares and triangles (see Figure 3) for five \nruns starting from different  random initial weights,  using a gradient-based learning \nmethod.  For ten successive frames, the same object would appear, but with the cen(cid:173)\ntre  varying randomly within  the central six-by-six patch of pixels.  In  the last  two \nframes, another randomly selected object would appear, so that object trajectories \noverlapped  by  two frames.  These  images  are  meant  to  be  a  crude approximation \nto what a  moving observer would see  while  watching multiple  moving objects in  a \nscene;  at  any  given  time  a  single  object  would  be  approximately centered  on  the \nretina but its exact location would always  be jittering from  moment  to moment. \nIn  these  simulations,  the  prior  distribution  for  the  temporal  covariances  between \nthe  two  groups  of units'  outputs  was  a  block-diagonal  configuration  as  in  Figure \n1 c),  but  with  three five-by-five  blocks  along the  diagonal.  Our choice  of a  block(cid:173)\ndiagonal prior distribution with three blocks encodes the constraint that units in  a \ngiven  block  in  one group  should  try to  agree  with  units  in  the  same  block  in  the \nother group; so each group should discover three classes of features.  The number of \nunits within a block was varied in preliminary experiments, and five units was found \nto  be  a  reasonable  number  to capture all  instances of each  object  class  (although \nthe performance of the algorithm seemed to be robust with respect to the number of \nunits  per block).  The learning took about  3000 iterations of steepest  descent  with \n\n\f938 \n\nSuzanna Becker \n\n\u2022 \n\n. - - - --\n.... . .. \n. . .  . . . . . ..  .... \n. .. \n.. .  I' ..\u2022. \n. ... ... \n... \n.. .  . ... \n. ...... \n.. . .. \n.\u2022\u2022 ..  ....... \n.\u2022 \n.\u2022\u2022 \n.\u2022 -1-\u00b7 \n. ...  .... . \n. . \n..  . .. ,'. ... \n.... .. \n~~;.  i1.=::::.:  ;~. :~:\"\n\u2022  1  \u2022 \u2022 \u2022  1: . . . . .   -.1 I '  \n\u2022 \u00b7 . .. \n- - - -,-\n. .. .. \n..... .' \n.. . \n..... \n... . ; \n\u2022\u2022. \n.. \n.1. \n. \n. ....... \u2022\u2022\u2022\u2022\u2022 II \n....  .... \n\u2022 1\n\u2022  1 \n0 ;\n\n.. \n....  .....  . .\u2022. . . ........  . ..... -i\u00b7\" . .. .  .. ..... \n\n' .. \n\u00b71 \u2022\u2022\u2022\u2022\u2022\u2022\u2022\u2022\u2022\u2022 ~  .\u2022\u2022..\u2022\u2022\u2022\u2022\u2022 :.. \n\n. .. \n.. \n... \n. ... \n:: \n\n. . . . . . .  \n\n' ~:.  ~::.... \n\n.  \u2022\n. \n. \u2022\u2022\u2022 \n\n........ \u2022\u2022\u2022\u2022\u2022\u2022 \n\n'a_\" _\"  \u2022\u2022 ..  ... \n\n:1\" .. \n\n..  ... .. I -\n\n. \u2022\u2022 \n\n, . .. . .  e o _ \n\n\u2022\u2022 . \u2022 \n\n. \u2022\u2022 \n\n..  .... \n\n\u2022  \u2022 \u2022 \u2022  \n\n\u2022 \u2022 \u2022  \n\n_  _  _ \n\n\u2022 \u2022  \n\n\u2022  .. \n\n.  _  _ \n\n__ \n\n.. \n\n.. \n\nI \n\n\u2022 \n\n.. \n\n.. \n\n\u2022 \n\u2022 \n' \n\u2022 \n\" \n~ . .  ~i~\"  ..\u2022 ::: \") ;: \n\n.  . \n\n1 ' 1 \n\u2022  \u2022 \u2022  ' \n':. \n\u2022  .. \n\u2022 \u2022   ..  i \" \n' ;' \n._\n:': \n\n.a \n\n' . \n\n, \n\n\u2022 \n\n.. \n\n\u00b7i . . . . . . . . . . .   : \" \n\n\u2022\u2022\u2022...\u2022 ~.~.: i\n.~:~ji----.. ......... \u00b7: .......... \n::: \n...... \n\u00b7 .... . \n\u00b7 ..... . \n\n\u2022  \u2022 \u2022  : . . .  \n. \n.; \n.... \n.... \n. . . .  .. . . ........ . \n. ..... . \n\nI  .. . \n.... \n.... \nI\u00b7 .. \n... \n.\u2022. \n. ..... \n\u2022  \u2022 \u2022   i....... \n..  \u2022  .. \n. . \n:::'  \"::  \\~. \n\n.. , \n. .. . \n.. ...... .. \n\ni\u00b7\u00b7  .. \u00b7\u00b7.. \n..... \nI ..  \" \n\n..  .... .. \n\u2022  ....... . \n\n..... \n\u2022  \u2022  \u2022 \n\n.;~ \n\n\u2022 \u2022  a  \u2022 \u2022 \u2022 \u2022  \n\n\u2022  .. \n\na  \u2022 \n:-\n\na \n\n\u2022 \n\n\u2022 \n\n\u2022 \n\n.\n\nFigure  4:  Weights  learned  by  one  of the  two  groups  of 15  units  in  the  first  layer. \nWhite  weights  are  positive  and  black  are  negative. \n\nmomentum  to  converge,  but  after  about  1000  iterations  only  very  small  weight \nchanges were made. \nWeights learned on a typical run for one of the two groups of fifteen  units are shown \nin  Figure 4.  The units'  weights  are displayed in  three rows  corresponding to units \nin  the three blocks in  the block-diagonal joint prior matrix.  Units within  the same \nblock  each  learned  different  instances  of the  same  pattern class.  For example,  on \nthis run units in  the first  block learned to detect circles in specific positions.  Units \nin  the  second  block  tended  to  learn  combinations  of either  horizontal  or  vertical \nlines,  or  sometimes  mixtures of the two.  In  the third block,  units learned  blurred, \nroughly  triangular  shape  detectors,  which  for  this  training  set  were  adequate  to \nrespond specifically to triangles. In all five runs the network converged to equivalent \nsolutions (only the groups' particular shape preferences varied across runs). \nVarying the number of units  per block from  three to five  (Le.  three three-by-three \nblocks  versus  three  five-by-five  blocks  of  units)  produced  similar  results,  except \nthat  with fewer  units  per block,  each  unit  tended  to capture multiple instances of \na  particular object  class in different  positions. \nA second layer of two groups of three units was added to the network, as  shown in \nFigure 2 b).  While keeping the first layer of weights frozen, this network was trained \nusing exactly the same cost function  as the first  layer for about 30 iterations using \na  gradient-based  learning  method.  This  time  the  prior joint  distribution  for  the \ntwo  classifications  was  a  three-by-three  matrix with  80%  of the density  along the \ndiagonal and 20% evenly distributed across the remainder of the distribution.  Units \nin  this  layer  learned  to  discriminate  fairly  well  between  the  three  object  classes, \nas  shown  in  Figure  5  a).  On  a  test  set  with  the  ambiguous  patterns  removed \n(Le.,  patterns containing multiple objects),  units in  the second layer achieved very \n\n\fJPMAX \n\na) \n\n1. \n\n~ 0.80 . \n\n0.50 \n\n. ~ \n\nc \n\n939 \n\nLoyer  2  Unit  Responses  on  Trai ning  Set \n\nI Circles \n\nSquares \nTriangles \n\n~ 1 \n'0  0.40 \nc i \u00a3 0.20 \n\n0.00 \n\n1.00 \n\nb) \n\nLoyer  2  Unit  Responses  on  Test  Set  (no  overlaps) \n\nc \n\n0 .. \n0 . \n\n:;;  0.80 \n\n. ;  0.60 \n\ni \n'0  0. 40 \nc \n0 \n\n0 1\u00b0,20 \n\nLIllI \n\n0.00 \n\n\u2022  \u2022 \n\nUM3 \n\n....  ----. \n\nUnit  6 \n\nFigure 5:  Response probabilities for  the  six output units to  each  of the  three  shapes. \n\naccurate object  discrimination as shown in  Figure 5 b). \nOn ambiguous patterns containing multiple objects, the network's performance was \ndisappointing.  The output units  would  sometimes produce the  \"correct\"  response, \ni.e.,  all  the  units  representing  the  shapes  present  in  the  image  would  be  partially \nactive.  Most often, however, only one of the correct shapes would be detected,  and \noccasionally  the  network's  response indicated  the  wrong  shape altogether.  It was \nhoped\u00b7 that the diagonally dominant prior mixed with a uniform density would allow \nunits  to occasionally disagree,  and they would  therefore  be able to represent  cases \nof multiple objects.  It may  have helped  to use a  similar prior for  the hidden  layer; \nhowever,  this would  increase the complexity of the learning considerably. \n\n3  DISCUSSION \n\nWe  have shown that the algorithm can  learn  2D-translation-invariant shape recog(cid:173)\nnition,  but  it  should  handle  equally  well  other  types  of transformations,  such  as \nrotation,  scaling  or  even  non-linear  transformations.  In  principle,  the  algorithm \nshould be applicable to real moving images;  this is currently being investigated.  Al(cid:173)\nthough we  have focused  here on  the temporal coherence constraint, the  algorithm \ncould  be  applied  equally  well  using  other  types  of  coherence,  such  as  coherence \n\n\f940 \n\nSuzanna Becker \n\nacross space or across different sensory modalities. \nNote  that  the  units  in  the first  layer of the network  did  not  learn  anything about \nthe geometric transformations between translated versions of the same object; they \nsimply  learned  to associate different  views  together.  In  this  respect,  the represen(cid:173)\ntation  learned  at  the  hidden  layer  is  similar  to  that  predicted  by  the  \"privileged \nviews\"  theory of viewpoint-invariant object  recognition advocated by  Weinshall  et \nal.  (1990)  (and others).  Their algorithm learns a similar representation in  a  single \nlayer of competing units with temporal Hebbian learning applied to the lateral con(cid:173)\nnections  between  these  units.  However,  the algorithm  proposed  here  goes  further \nin  that it  can  be applied  to subsequent  stages  of learning to discover  higher-order \nobject classes. \nYuille  et  al.  (1994)  have  previously  proposed  an  algorithm  based  on  similar  prin(cid:173)\nciples,  which  also  involves  maximizing the  log  probability of the  network outputs \nunder a prior; in one special case it is equivalent to Becker and Hinton's Imax algo(cid:173)\nrithm.  The algorithm proposed here differs substantially, in that we are dealing with \nthe ensemble-averaged joint probabilities of two populations of units, and fitting this \nquantity to a prior; further,  Yuille  et aI's scheme employs back-propagation. \nOne  challenge  for  future  work  is  to  train  a  network  with  smaller  receptive  fields \nfor  the first  layer units,  on images of objects with  common low-level features,  such \nas  squares  and  rectangles.  At  least  three  layers  of weights  would  be  required  to \nsolve this task:  units in  the first  layer would have to learn local object parts such as \ncorners, while units in the next layer could group parts into viewpoint-specific whole \nobjects and in  the top layer viewpoint-invariance, in  principle,  could be achieved. \n\nAcknowledgements \n\nHelpful  comments from  Geoff Hinton, Peter Dayan, Tony Plate and Chris Williams \nare gratefully acknowledged. \n\nReferences \n\nBecker,  S.  (1993).  Learning to categorize objects  using  temporal  coherence.  In \nAdvances  in  Neural  Information  Processing  Systems  5,  pages  361- 368. Morgan \nKaufmann. \nBecker,  S.  and Hinton,  G.  E.  (1992).  A self-organizing neural  network that dis(cid:173)\ncovers surfaces in  random-dot stereograms.  Nature,  355: 161-163. \nde  Sa,  V.  R.  (1993).  Minimizing  disagreement  for  self-supervised  classification. \nIn  Proceedings  of the  1993 Connectionist Models  Summer School,  pages 300-307. \nLawrence Erlbaum associates. \nF51diak,  P.  (1991).  Learning invariance from  transformation  sequences.  Neural \nComputation,  3(2):194-200. \nWeinshall,  D.,  Edelman,  S.,  and  Biilthoff,  H.  H.  (1990).  A  self-organizing \nmultiple-view  representation of 3D  objects.  In  Advances  in  Neural  Information \nProcessing  Systems 2,  pages 274-282.  Morgan  Kaufmann. \nYuille,  A.  L.,  Stelios,  M.  S.,  and  Xu,  L.  (1994).  Bayesian  Self-Organization. \nTechnical Report No.  92-10, Harvard Robotics Laboratory. \n\n\f", "award": [], "sourceid": 894, "authors": [{"given_name": "Suzanna", "family_name": "Becker", "institution": null}]}