 ac5013a026
			
		
	
	
		ac5013a026
		
	
	
	
	
		
			
			git-svn-id: svn://svn.cy55.de/Zope3/src/cybertools/trunk@2585 fd906abe-77d9-0310-91a1-e0d9ade77398
		
			
				
	
	
		
			92 lines
		
	
	
	
		
			2.2 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			92 lines
		
	
	
	
		
			2.2 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| #
 | |
| #  Copyright (c) 2008 Helmut Merz helmutm@cy55.de
 | |
| #
 | |
| #  This program is free software; you can redistribute it and/or modify
 | |
| #  it under the terms of the GNU General Public License as published by
 | |
| #  the Free Software Foundation; either version 2 of the License, or
 | |
| #  (at your option) any later version.
 | |
| #
 | |
| #  This program is distributed in the hope that it will be useful,
 | |
| #  but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| #  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
| #  GNU General Public License for more details.
 | |
| #
 | |
| #  You should have received a copy of the GNU General Public License
 | |
| #  along with this program; if not, write to the Free Software
 | |
| #  Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
 | |
| #
 | |
| 
 | |
| """
 | |
| Crawl base and sample classes.
 | |
| 
 | |
| $Id$
 | |
| """
 | |
| 
 | |
| from zope.interface import implements
 | |
| 
 | |
| from cybertools.agent.base.agent import Master
 | |
| from cybertools.agent.core.agent import QueueableAgent
 | |
| from cybertools.agent.interfaces import ICrawler
 | |
| from cybertools.agent.interfaces import IResource, IMetadataSet
 | |
| from cybertools.agent.components import agents
 | |
| from twisted.internet.defer import succeed
 | |
| 
 | |
| 
 | |
| class Crawler(QueueableAgent):
 | |
| 
 | |
|     implements(ICrawler)
 | |
| 
 | |
|     def __init__(self, master, params={}):
 | |
|         super(Crawler, self).__init__(master)
 | |
| 
 | |
|     def process(self):
 | |
|         return self.collect()
 | |
| 
 | |
|     def collect(self, filter=None):
 | |
|         d = defer.succeed([])
 | |
|         return d
 | |
| 
 | |
| 
 | |
| class SampleCrawler(Crawler):
 | |
| 
 | |
|     def collect(self, filter=None):
 | |
|         print 'SampleCrawler is collecting.'
 | |
|         d = succeed([])
 | |
|         return d
 | |
| 
 | |
| agents.register(SampleCrawler, Master, name='crawl.sample')
 | |
| 
 | |
| 
 | |
| class Resource(object):
 | |
| 
 | |
|     implements(IResource)
 | |
| 
 | |
|     data = file = path = None
 | |
|     type = 'sample'
 | |
|     contentType = 'text/plain'
 | |
|     encoding = ''
 | |
|     application = 'sample'
 | |
|     metadata = None
 | |
| 
 | |
|     def __init__(self, data=None, **kw):
 | |
|         if data is not None:
 | |
|             self.data = data
 | |
|         for k, v in kw.items():
 | |
|             setattr(self, k, v)
 | |
|         self.subResources = []
 | |
| 
 | |
| 
 | |
| class Metadata(dict):
 | |
| 
 | |
|     implements(IMetadataSet)
 | |
| 
 | |
|     def __init__(self, data=dict()):
 | |
|         self.update(data)
 | |
| 
 | |
|     def asXML(self):
 | |
|         # TODO...
 | |
|         return ''
 | |
| 
 | |
|     def set(self, key, value):
 | |
|         self['key'] = value
 | |
| 
 |