00001 /*!@file Channels/MotionChannel.C */ 00002 00003 // //////////////////////////////////////////////////////////////////// // 00004 // The iLab Neuromorphic Vision C++ Toolkit - Copyright (C) 2000-2005 // 00005 // by the University of Southern California (USC) and the iLab at USC. // 00006 // See http://iLab.usc.edu for information about this project. // 00007 // //////////////////////////////////////////////////////////////////// // 00008 // Major portions of the iLab Neuromorphic Vision Toolkit are protected // 00009 // under the U.S. patent ``Computation of Intrinsic Perceptual Saliency // 00010 // in Visual Environments, and Applications'' by Christof Koch and // 00011 // Laurent Itti, California Institute of Technology, 2001 (patent // 00012 // pending; application number 09/912,225 filed July 23, 2001; see // 00013 // http://pair.uspto.gov/cgi-bin/final/home.pl for current status). // 00014 // //////////////////////////////////////////////////////////////////// // 00015 // This file is part of the iLab Neuromorphic Vision C++ Toolkit. // 00016 // // 00017 // The iLab Neuromorphic Vision C++ Toolkit is free software; you can // 00018 // redistribute it and/or modify it under the terms of the GNU General // 00019 // Public License as published by the Free Software Foundation; either // 00020 // version 2 of the License, or (at your option) any later version. // 00021 // // 00022 // The iLab Neuromorphic Vision C++ Toolkit is distributed in the hope // 00023 // that it will be useful, but WITHOUT ANY WARRANTY; without even the // 00024 // implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR // 00025 // PURPOSE. See the GNU General Public License for more details. // 00026 // // 00027 // You should have received a copy of the GNU General Public License // 00028 // along with the iLab Neuromorphic Vision C++ Toolkit; if not, write // 00029 // to the Free Software Foundation, Inc., 59 Temple Place, Suite 330, // 00030 // Boston, MA 02111-1307 USA. // 00031 // //////////////////////////////////////////////////////////////////// // 00032 // 00033 // Primary maintainer for this file: 00034 // $HeadURL: svn://isvn.usc.edu/software/invt/trunk/saliency/src/Channels/MotionChannel.C $ 00035 // $Id: MotionChannel.C 11208 2009-05-20 02:03:21Z itti $ 00036 // 00037 00038 #ifndef MOTIONCHANNEL_C_DEFINED 00039 #define MOTIONCHANNEL_C_DEFINED 00040 00041 #include "Channels/MotionChannel.H" 00042 00043 #include "Channels/ChannelOpts.H" 00044 #include "Channels/DirectionChannel.H" 00045 #include "Component/OptionManager.H" 00046 #include "rutz/trace.h" 00047 00048 // ###################################################################### 00049 // MotionChannel member definitions: 00050 // ###################################################################### 00051 00052 // ###################################################################### 00053 MotionChannel::MotionChannel(OptionManager& mgr) : 00054 ComplexChannel(mgr, "Motion", "motion", MOTION), 00055 itsPyrType("MotionChannelPyramidType", this, Gaussian5), 00056 itsNumDirs(&OPT_NumDirections, this) // see Channels/ChannelOpts.{H,C} 00057 { 00058 GVX_TRACE(__PRETTY_FUNCTION__); 00059 // let's create our subchannels (may be reconfigured later if our 00060 // number of directions changes): 00061 buildSubChans(); 00062 } 00063 00064 // ###################################################################### 00065 MotionChannel::~MotionChannel() 00066 { 00067 GVX_TRACE(__PRETTY_FUNCTION__); 00068 } 00069 00070 // ###################################################################### 00071 DirectionChannel& MotionChannel::dirChan(const uint idx) const 00072 { 00073 GVX_TRACE(__PRETTY_FUNCTION__); 00074 return *(dynCast<DirectionChannel>(subChan(idx))); 00075 } 00076 00077 // ###################################################################### 00078 void MotionChannel::buildSubChans() 00079 { 00080 GVX_TRACE(__PRETTY_FUNCTION__); 00081 // kill any subchans we may have had... 00082 this->removeAllSubChans(); 00083 00084 // let's instantiate our subchannels now that we know how many we 00085 // want. They will inherit the current values (typically 00086 // post-command-line parsing) of all their options as they are 00087 // constructed: 00088 LINFO("Using %d directions spanning [0..360]deg", itsNumDirs.getVal()); 00089 for (uint i = 0; i < itsNumDirs.getVal(); ++i) 00090 { 00091 nub::ref<DirectionChannel> chan = 00092 makeSharedComp 00093 (new DirectionChannel(getManager(), i, 00094 360.0 * double(i) / 00095 double(itsNumDirs.getVal()), 00096 itsPyrType.getVal())); 00097 this->addSubChan(chan); 00098 00099 chan->exportOptions(MC_RECURSE); 00100 } 00101 } 00102 00103 // ###################################################################### 00104 void MotionChannel::paramChanged(ModelParamBase* const param, 00105 const bool valueChanged, 00106 ParamClient::ChangeStatus* status) 00107 { 00108 GVX_TRACE(__PRETTY_FUNCTION__); 00109 ComplexChannel::paramChanged(param, valueChanged, status); 00110 00111 // if the param is our number of orientations and it has become 00112 // different from our number of channels, let's reconfigure: 00113 if (param == &itsNumDirs && 00114 numChans() != itsNumDirs.getVal()) 00115 buildSubChans(); 00116 } 00117 00118 // ###################################################################### 00119 void MotionChannel::doInput(const InputFrame& inframe) 00120 { 00121 GVX_TRACE(__PRETTY_FUNCTION__); 00122 ASSERT(inframe.grayFloat().initialized()); 00123 00124 // compute Reichardt motion detection into several directions 00125 for (uint dir = 0; dir < numChans(); ++dir) 00126 { 00127 subChan(dir)->input(inframe); 00128 LINFO("Motion pyramid (%d/%d) ok.", dir+1, numChans()); 00129 } 00130 } 00131 00132 // ###################################################################### 00133 /* So things look consistent in everyone's emacs... */ 00134 /* Local Variables: */ 00135 /* indent-tabs-mode: nil */ 00136 /* End: */ 00137 00138 #endif // MOTIONCHANNEL_C_DEFINED