rem
stringlengths
1
322k
add
stringlengths
0
2.05M
context
stringlengths
4
228k
meta
stringlengths
156
215
opts['conditions'] = {'method':meth}
opts['conditions'] = {'method':[meth]}
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
da7004651758a3ba73f50c2731e2b08ce10f1730 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/da7004651758a3ba73f50c2731e2b08ce10f1730/base.py
self.connect("formatted_" + route_name, "%s.:(format);%s")
self.connect("formatted_" + route_name, "%s\.:(format);%s")
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
da7004651758a3ba73f50c2731e2b08ce10f1730 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/da7004651758a3ba73f50c2731e2b08ce10f1730/base.py
self.connect("%s.:(format)" % collection_path, **route_options)
self.connect("%s\.:(format)" % collection_path, **route_options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
da7004651758a3ba73f50c2731e2b08ce10f1730 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/da7004651758a3ba73f50c2731e2b08ce10f1730/base.py
conditions={'method':'GET'}, **options)
conditions={'method':['GET']}, **options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
da7004651758a3ba73f50c2731e2b08ce10f1730 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/da7004651758a3ba73f50c2731e2b08ce10f1730/base.py
collection_path + ".:(format)", action='index', conditions={'method':'GET'}, **options)
collection_path + "\.:(format)", action='index', conditions={'method':['GET']}, **options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
da7004651758a3ba73f50c2731e2b08ce10f1730 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/da7004651758a3ba73f50c2731e2b08ce10f1730/base.py
path = (action == 'new' and action + '.:(format)') or \ "%s.:(format);%s" % (new_path, action)
path = (action == 'new' and new_path + '\.:(format)') or \ "%s\.:(format);%s" % (new_path, action)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
da7004651758a3ba73f50c2731e2b08ce10f1730 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/da7004651758a3ba73f50c2731e2b08ce10f1730/base.py
"%s.:(format);%s" % (member_path, action), **route_options)
"%s\.:(format);%s" % (member_path, action), **route_options)
def requirements_for(meth): opts = options.copy() if method != 'any': opts['conditions'] = {'method':meth} return opts
da7004651758a3ba73f50c2731e2b08ce10f1730 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/da7004651758a3ba73f50c2731e2b08ce10f1730/base.py
hostmatch = config.environ['HTTP_HOST'].split(':')
fullhost = config.environ.get('HTTP_HOST') or config.environ.get('SERVER_NAME') hostmatch = fullhost.split(':')
def _subdomain_check(config, kargs): if config.mapper.sub_domains: subdomain = kargs.pop('sub_domain', None) hostmatch = config.environ['HTTP_HOST'].split(':') host = hostmatch[0] port = '' if len(hostmatch) > 1: port += ':' + hostmatch[1] sub_match = re.compile('^.+?\.(%s)$' % config.mapper.domain_match) domain = re.sub(sub_match, r'\1', host) if subdomain and not host.startswith(subdomain) and \ subdomain not in config.mapper.sub_domains_ignore: kargs['_host'] = subdomain + '.' + domain + port elif (subdomain in config.mapper.sub_domains_ignore or subdomain is None) and domain != host: kargs['_host'] = domain + port return kargs else: return kargs
dde994943413b0ff3e99dac7b8d91fd338ed23c7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/dde994943413b0ff3e99dac7b8d91fd338ed23c7/util.py
controllers.extend(find_controllers(filename, prefix=fname+'/'))
controllers.extend(find_controllers(filename, prefix=prefix+fname+'/'))
def find_controllers(dirname, prefix=''): controllers = [] for fname in os.listdir(dirname): filename = dirname + '/' + fname if os.path.isfile(filename) and re.match('^[^_]{1,1}.*\.py$', fname): controllers.append(prefix + fname[:-3]) elif os.path.isdir(filename): controllers.extend(find_controllers(filename, prefix=fname+'/')) return controllers
9a4ac958a07c4ddd56432c85cc5d63ecd4d631f2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/9a4ac958a07c4ddd56432c85cc5d63ecd4d631f2/util.py
class RouteConfig(threadinglocal.local): pass
def keysort(a, b): am = a.minkeys a = a.maxkeys b = b.maxkeys lendiffa = len(keys^a) lendiffb = len(keys^b) # If they both match, don't switch them if lendiffa == 0 and lendiffb == 0: return 0 # First, if a matches exactly, use it if lendiffa == 0: return -1 # Or b matches exactly, use it if lendiffb == 0: return 1 # Neither matches exactly, return the one with the most in common if cmp(lendiffa,lendiffb) != 0: return cmp(lendiffa,lendiffb) # Neither matches exactly, but if they both have just as much in common if len(keys&b) == len(keys&a): return cmp(len(a),len(b)) # Then we return the shortest of the two # Otherwise, we return the one that has the most in common else: return cmp(len(keys&b), len(keys&a))
7bfa64f1dbabfd9ab39e416435bdda59b9db811c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/7bfa64f1dbabfd9ab39e416435bdda59b9db811c/base.py
_shared_state = RouteConfig()
__shared_state = threadinglocal.local()
def keysort(a, b): am = a.minkeys a = a.maxkeys b = b.maxkeys lendiffa = len(keys^a) lendiffb = len(keys^b) # If they both match, don't switch them if lendiffa == 0 and lendiffb == 0: return 0 # First, if a matches exactly, use it if lendiffa == 0: return -1 # Or b matches exactly, use it if lendiffb == 0: return 1 # Neither matches exactly, return the one with the most in common if cmp(lendiffa,lendiffb) != 0: return cmp(lendiffa,lendiffb) # Neither matches exactly, but if they both have just as much in common if len(keys&b) == len(keys&a): return cmp(len(a),len(b)) # Then we return the shortest of the two # Otherwise, we return the one that has the most in common else: return cmp(len(keys&b), len(keys&a))
7bfa64f1dbabfd9ab39e416435bdda59b9db811c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/7bfa64f1dbabfd9ab39e416435bdda59b9db811c/base.py
return self._shared_state.__getattr__(name)
return self.__shared_state.__getattr__(name)
def __getattr__(self, name): return self._shared_state.__getattr__(name)
7bfa64f1dbabfd9ab39e416435bdda59b9db811c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/7bfa64f1dbabfd9ab39e416435bdda59b9db811c/base.py
return self._shared_state.__setattr__(name, value)
return self.__shared_state.__setattr__(name, value)
def __setattr__(self, name, value): return self._shared_state.__setattr__(name, value)
7bfa64f1dbabfd9ab39e416435bdda59b9db811c /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/7bfa64f1dbabfd9ab39e416435bdda59b9db811c/base.py
newargs = route.defaults.copy() host = newargs.pop('host', '') protocol = newargs.pop('protocol', '') del newargs['_static'] newargs.update(kargs) url = route.generate(_ignore_req_list=True, **newargs) if not url: url = '' kargs = None
url = route.routepath
def url_for(*args, **kargs): """Generates a URL All keys given to url_for are sent to the Routes Mapper instance for generation except for:: anchor specified the anchor name to be appened to the path host overrides the default (current) host if provided protocol overrides the default (current) protocol if provided The URL is generated based on the rest of the keys. When generating a new URL, values will be used from the current request's parameters (if present). The following rules are used to determine when and how to keep the current requests parameters: * If the controller is present and begins with '/', no defaults are used * If the controller is changed, action is set to 'index' unless otherwise specified For example, if the current request yielded a dict of {'controller': 'blog', 'action': 'view', 'id': 2}, with the standard ':controller/:action/:id' route, you'd get the following results:: url_for(id=4) => '/blog/view/4', url_for(controller='/admin') => '/admin', url_for(controller='admin') => '/admin/index/4' url_for(action='edit') => '/blog/post/4', url_for(action='list', id=None) => '/blog/list' **Static and Named Routes** If there is a string present as the first argument, a lookup is done against the named routes table to see if there's any matching routes. The keyword defaults used with static routes will be sent in as GET query arg's if a route matches. If no route by that name is found, the string is assumed to be a raw URL. Should the raw URL begin with ``/`` then appropriate SCRIPT_NAME data will be added if present, otherwise the string will be used as the url with keyword args becoming GET query args. """ anchor = kargs.get('anchor') host = kargs.get('host') protocol = kargs.get('protocol') # Remove special words from kargs, convert placeholders for key in ['anchor', 'host', 'protocol']: if kargs.get(key): del kargs[key] if kargs.has_key(key+'_'): kargs[key] = kargs[key+'_'] del kargs[key+'_'] config = request_config() route = None static = False url = '' if len(args) > 0: route = config.mapper._routenames.get(args[0]) if route and route.defaults.has_key('_static'): static = True newargs = route.defaults.copy() host = newargs.pop('host', '') protocol = newargs.pop('protocol', '') del newargs['_static'] newargs.update(kargs) url = route.generate(_ignore_req_list=True, **newargs) if not url: url = '' kargs = None #url = route.routepath # No named route found, assume the argument is a relative path if not route: static = True url = args[0] if url.startswith('/') and hasattr(config, 'environ') \ and config.environ.get('SCRIPT_NAME'): url = config.environ.get('SCRIPT_NAME') + url if static: if kargs: url += '?' url += urllib.urlencode(kargs) if not static: if route: newargs = route.defaults.copy() newargs.update(kargs) # If this route has a filter, apply it if route.filter: newargs = route.filter(newargs) else: newargs = _screenargs(kargs) url = config.mapper.generate(**newargs) if config.mapper.append_slash and not url.endswith('/'): url += '/' if anchor: url += '#' + _url_quote(anchor) if host or protocol: if not host: host = config.host if not protocol: protocol = config.protocol url = protocol + '://' + host + url return url
892bb7149113b822705b8b36221931c8711626e7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/892bb7149113b822705b8b36221931c8711626e7/util.py
if isinsance(use_sd, list) and sub_domain not in use_sd:
if isinstance(use_sd, list) and sub_domain not in use_sd:
def match(self, url, environ=None, sub_domains=False, sub_domains_ignore=None, domain_match=''): """Match a url to our regexp. While the regexp might match, this operation isn't guaranteed as there's other factors that can cause a match to fail even though the regexp succeeds (Default that was relied on wasn't given, requirement regexp doesn't pass, etc.). Therefore the calling function shouldn't assume this will return a valid dict, the other possible return is False if a match doesn't work out. """ # Static routes don't match, they generate only if self.static: return False if url.endswith('/') and len(url) > 1: url = url[:-1] m = self.regmatch.match(url) if not m: return False if not environ: environ = {} sub_domain = None if environ.get('HTTP_HOST') and sub_domains: host = environ['HTTP_HOST'].split(':')[0] sub_match = re.compile('^(.+?)\.%s$' % domain_match) subdomain = re.sub(sub_match, r'\1', host) if subdomain not in sub_domains_ignore and host != subdomain: sub_domain = subdomain if self.conditions: if self.conditions.has_key('method') and \ environ.get('REQUEST_METHOD') not in self.conditions['method']: return False # Check sub-domains? use_sd = self.conditions.get('sub_domain') if use_sd and not sub_domain: return False if isinsance(use_sd, list) and sub_domain not in use_sd: return False matchdict = m.groupdict() result = {} extras = frozenset(self.defaults.keys()) - frozenset(matchdict.keys()) for key,val in matchdict.iteritems(): if not val and self.defaults.has_key(key) and self.defaults[key]: result[key] = self.defaults[key] else: result[key] = val for key in extras: result[key] = self.defaults[key] # If there's a function, call it with environ and expire if it # returns False if self.conditions and self.conditions.has_key('function') and \ not self.conditions['function'](environ, result): return False if sub_domain: result['sub_domain'] = sub_domain return result
be5725db10acbd057f664fa467379bd9aba85382 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/be5725db10acbd057f664fa467379bd9aba85382/base.py
var = part['name']
var = re.escape(part['name'])
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the remainder has either a string match, or a non-defaulted # regexp match on a key, allblank remembers if the rest could possible be completely empty (rest, noreqs, allblank) = ('', True, True) if len(path[1:]) > 0: (rest, noreqs, allblank) = self.buildnextreg(path[1:], clist) if isinstance(part, dict) and part['type'] == ':': var = part['name'] partreg = '' # First we plug in the proper part matcher if self.reqs.has_key(var): partreg = '(?P<' + var + '>' + self.reqs[var] + ')' elif var == 'controller': partreg = '(?P<' + var + '>' + '|'.join(clist) + ')' else: if len(path) > 1: partreg = '(?P<' + var + '>[^' + path[1][0] +']+)' else: partreg = '(?P<' + var + '>[^/]+)' if self.reqs.has_key(var): noreqs = False if not self.defaults.has_key(var): allblank = False noreqs = False # Now we determine if its optional, or required. This changes depending on what is in # the rest of the match. If noreqs is true, then its possible the entire thing is optional # as there's no reqs or string matches. if noreqs: # The rest is optional, but now we have an optional with a regexp. Wrap to ensure that if we match # anything, we match our regexp first. It's still possible we could be completely blank as we have # a default if self.reqs.has_key(var) and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Or we have a regexp match with no default, so now being completely blank form here on out isn't # possible elif self.reqs.has_key(var): allblank = False reg = partreg + rest # Or we have a default with no regexp, don't touch the allblank elif self.defaults.has_key(var): reg = partreg + '?' + rest # Or we have a key with no default, and no reqs. Not possible to be all blank from here else: allblank = False reg = partreg + rest # In this case, we have something dangling that might need to be matched else: # If they can all be blank, and we have a default here, we know its # safe to make everything from here optional. Since something else in # the chain does have req's though, we have to make the partreg here # required to continue matching if allblank and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Same as before, but they can't all be blank, so we have to require it all to ensure # our matches line up right else: reg = partreg + rest elif isinstance(part, dict) and part['type'] == '*': var = part['name'] if noreqs: if self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: reg = '(?P<' + var + '>.*)' + rest allblank = False noreqs = False else: if allblank and self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest elif self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: allblank = False noreqs = False reg = '(?P<' + var + '>.*)' + rest elif part.endswith('/'): if allblank: reg = part[:-1] + '(/' + rest + ')?' else: allblank = False reg = part + rest # We have a normal string here, this is a req, and it prevents us from being all blank else: noreqs = False allblank = False reg = part + rest return (reg, noreqs, allblank)
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
partreg = '(?P<' + var + '>' + '|'.join(clist) + ')'
partreg = '(?P<' + var + '>' + '|'.join(map(re.escape, clist)) + ')'
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the remainder has either a string match, or a non-defaulted # regexp match on a key, allblank remembers if the rest could possible be completely empty (rest, noreqs, allblank) = ('', True, True) if len(path[1:]) > 0: (rest, noreqs, allblank) = self.buildnextreg(path[1:], clist) if isinstance(part, dict) and part['type'] == ':': var = part['name'] partreg = '' # First we plug in the proper part matcher if self.reqs.has_key(var): partreg = '(?P<' + var + '>' + self.reqs[var] + ')' elif var == 'controller': partreg = '(?P<' + var + '>' + '|'.join(clist) + ')' else: if len(path) > 1: partreg = '(?P<' + var + '>[^' + path[1][0] +']+)' else: partreg = '(?P<' + var + '>[^/]+)' if self.reqs.has_key(var): noreqs = False if not self.defaults.has_key(var): allblank = False noreqs = False # Now we determine if its optional, or required. This changes depending on what is in # the rest of the match. If noreqs is true, then its possible the entire thing is optional # as there's no reqs or string matches. if noreqs: # The rest is optional, but now we have an optional with a regexp. Wrap to ensure that if we match # anything, we match our regexp first. It's still possible we could be completely blank as we have # a default if self.reqs.has_key(var) and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Or we have a regexp match with no default, so now being completely blank form here on out isn't # possible elif self.reqs.has_key(var): allblank = False reg = partreg + rest # Or we have a default with no regexp, don't touch the allblank elif self.defaults.has_key(var): reg = partreg + '?' + rest # Or we have a key with no default, and no reqs. Not possible to be all blank from here else: allblank = False reg = partreg + rest # In this case, we have something dangling that might need to be matched else: # If they can all be blank, and we have a default here, we know its # safe to make everything from here optional. Since something else in # the chain does have req's though, we have to make the partreg here # required to continue matching if allblank and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Same as before, but they can't all be blank, so we have to require it all to ensure # our matches line up right else: reg = partreg + rest elif isinstance(part, dict) and part['type'] == '*': var = part['name'] if noreqs: if self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: reg = '(?P<' + var + '>.*)' + rest allblank = False noreqs = False else: if allblank and self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest elif self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: allblank = False noreqs = False reg = '(?P<' + var + '>.*)' + rest elif part.endswith('/'): if allblank: reg = part[:-1] + '(/' + rest + ')?' else: allblank = False reg = part + rest # We have a normal string here, this is a req, and it prevents us from being all blank else: noreqs = False allblank = False reg = part + rest return (reg, noreqs, allblank)
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
reg = part[:-1] + '(/' + rest + ')?'
reg = re.escape(part[:-1]) + '(/' + rest + ')?'
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the remainder has either a string match, or a non-defaulted # regexp match on a key, allblank remembers if the rest could possible be completely empty (rest, noreqs, allblank) = ('', True, True) if len(path[1:]) > 0: (rest, noreqs, allblank) = self.buildnextreg(path[1:], clist) if isinstance(part, dict) and part['type'] == ':': var = part['name'] partreg = '' # First we plug in the proper part matcher if self.reqs.has_key(var): partreg = '(?P<' + var + '>' + self.reqs[var] + ')' elif var == 'controller': partreg = '(?P<' + var + '>' + '|'.join(clist) + ')' else: if len(path) > 1: partreg = '(?P<' + var + '>[^' + path[1][0] +']+)' else: partreg = '(?P<' + var + '>[^/]+)' if self.reqs.has_key(var): noreqs = False if not self.defaults.has_key(var): allblank = False noreqs = False # Now we determine if its optional, or required. This changes depending on what is in # the rest of the match. If noreqs is true, then its possible the entire thing is optional # as there's no reqs or string matches. if noreqs: # The rest is optional, but now we have an optional with a regexp. Wrap to ensure that if we match # anything, we match our regexp first. It's still possible we could be completely blank as we have # a default if self.reqs.has_key(var) and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Or we have a regexp match with no default, so now being completely blank form here on out isn't # possible elif self.reqs.has_key(var): allblank = False reg = partreg + rest # Or we have a default with no regexp, don't touch the allblank elif self.defaults.has_key(var): reg = partreg + '?' + rest # Or we have a key with no default, and no reqs. Not possible to be all blank from here else: allblank = False reg = partreg + rest # In this case, we have something dangling that might need to be matched else: # If they can all be blank, and we have a default here, we know its # safe to make everything from here optional. Since something else in # the chain does have req's though, we have to make the partreg here # required to continue matching if allblank and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Same as before, but they can't all be blank, so we have to require it all to ensure # our matches line up right else: reg = partreg + rest elif isinstance(part, dict) and part['type'] == '*': var = part['name'] if noreqs: if self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: reg = '(?P<' + var + '>.*)' + rest allblank = False noreqs = False else: if allblank and self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest elif self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: allblank = False noreqs = False reg = '(?P<' + var + '>.*)' + rest elif part.endswith('/'): if allblank: reg = part[:-1] + '(/' + rest + ')?' else: allblank = False reg = part + rest # We have a normal string here, this is a req, and it prevents us from being all blank else: noreqs = False allblank = False reg = part + rest return (reg, noreqs, allblank)
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
reg = part + rest
reg = re.escape(part) + rest
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the remainder has either a string match, or a non-defaulted # regexp match on a key, allblank remembers if the rest could possible be completely empty (rest, noreqs, allblank) = ('', True, True) if len(path[1:]) > 0: (rest, noreqs, allblank) = self.buildnextreg(path[1:], clist) if isinstance(part, dict) and part['type'] == ':': var = part['name'] partreg = '' # First we plug in the proper part matcher if self.reqs.has_key(var): partreg = '(?P<' + var + '>' + self.reqs[var] + ')' elif var == 'controller': partreg = '(?P<' + var + '>' + '|'.join(clist) + ')' else: if len(path) > 1: partreg = '(?P<' + var + '>[^' + path[1][0] +']+)' else: partreg = '(?P<' + var + '>[^/]+)' if self.reqs.has_key(var): noreqs = False if not self.defaults.has_key(var): allblank = False noreqs = False # Now we determine if its optional, or required. This changes depending on what is in # the rest of the match. If noreqs is true, then its possible the entire thing is optional # as there's no reqs or string matches. if noreqs: # The rest is optional, but now we have an optional with a regexp. Wrap to ensure that if we match # anything, we match our regexp first. It's still possible we could be completely blank as we have # a default if self.reqs.has_key(var) and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Or we have a regexp match with no default, so now being completely blank form here on out isn't # possible elif self.reqs.has_key(var): allblank = False reg = partreg + rest # Or we have a default with no regexp, don't touch the allblank elif self.defaults.has_key(var): reg = partreg + '?' + rest # Or we have a key with no default, and no reqs. Not possible to be all blank from here else: allblank = False reg = partreg + rest # In this case, we have something dangling that might need to be matched else: # If they can all be blank, and we have a default here, we know its # safe to make everything from here optional. Since something else in # the chain does have req's though, we have to make the partreg here # required to continue matching if allblank and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Same as before, but they can't all be blank, so we have to require it all to ensure # our matches line up right else: reg = partreg + rest elif isinstance(part, dict) and part['type'] == '*': var = part['name'] if noreqs: if self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: reg = '(?P<' + var + '>.*)' + rest allblank = False noreqs = False else: if allblank and self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest elif self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: allblank = False noreqs = False reg = '(?P<' + var + '>.*)' + rest elif part.endswith('/'): if allblank: reg = part[:-1] + '(/' + rest + ')?' else: allblank = False reg = part + rest # We have a normal string here, this is a req, and it prevents us from being all blank else: noreqs = False allblank = False reg = part + rest return (reg, noreqs, allblank)
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
reg = part + rest
reg = re.escape(part) + rest
def buildnextreg(self, path, clist): """Recursively build our regexp given a path, and a controller list. Returns the regular expression string, and two booleans that can be ignored as they're only used internally by buildnextreg """ if path: part = path[0] else: part = '' reg = '' # noreqs will remember whether the remainder has either a string match, or a non-defaulted # regexp match on a key, allblank remembers if the rest could possible be completely empty (rest, noreqs, allblank) = ('', True, True) if len(path[1:]) > 0: (rest, noreqs, allblank) = self.buildnextreg(path[1:], clist) if isinstance(part, dict) and part['type'] == ':': var = part['name'] partreg = '' # First we plug in the proper part matcher if self.reqs.has_key(var): partreg = '(?P<' + var + '>' + self.reqs[var] + ')' elif var == 'controller': partreg = '(?P<' + var + '>' + '|'.join(clist) + ')' else: if len(path) > 1: partreg = '(?P<' + var + '>[^' + path[1][0] +']+)' else: partreg = '(?P<' + var + '>[^/]+)' if self.reqs.has_key(var): noreqs = False if not self.defaults.has_key(var): allblank = False noreqs = False # Now we determine if its optional, or required. This changes depending on what is in # the rest of the match. If noreqs is true, then its possible the entire thing is optional # as there's no reqs or string matches. if noreqs: # The rest is optional, but now we have an optional with a regexp. Wrap to ensure that if we match # anything, we match our regexp first. It's still possible we could be completely blank as we have # a default if self.reqs.has_key(var) and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Or we have a regexp match with no default, so now being completely blank form here on out isn't # possible elif self.reqs.has_key(var): allblank = False reg = partreg + rest # Or we have a default with no regexp, don't touch the allblank elif self.defaults.has_key(var): reg = partreg + '?' + rest # Or we have a key with no default, and no reqs. Not possible to be all blank from here else: allblank = False reg = partreg + rest # In this case, we have something dangling that might need to be matched else: # If they can all be blank, and we have a default here, we know its # safe to make everything from here optional. Since something else in # the chain does have req's though, we have to make the partreg here # required to continue matching if allblank and self.defaults.has_key(var): reg = '(' + partreg + rest + ')?' # Same as before, but they can't all be blank, so we have to require it all to ensure # our matches line up right else: reg = partreg + rest elif isinstance(part, dict) and part['type'] == '*': var = part['name'] if noreqs: if self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: reg = '(?P<' + var + '>.*)' + rest allblank = False noreqs = False else: if allblank and self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest elif self.defaults.has_key(var): reg = '(?P<' + var + '>.*)' + rest else: allblank = False noreqs = False reg = '(?P<' + var + '>.*)' + rest elif part.endswith('/'): if allblank: reg = part[:-1] + '(/' + rest + ')?' else: allblank = False reg = part + rest # We have a normal string here, this is a req, and it prevents us from being all blank else: noreqs = False allblank = False reg = part + rest return (reg, noreqs, allblank)
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
self.connect("formatted_" + route_name, "%s\.:(format);%s")
self.connect("formatted_" + route_name, "%s.:(format);%s")
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
self.connect("%s\.:(format)" % collection_path, **route_options)
self.connect("%s.:(format)" % collection_path, **route_options)
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
collection_path + "\.:(format)", action='index',
collection_path + ".:(format)", action='index',
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
path = (action == 'new' and new_path + '\.:(format)') or \ "%s\.:(format);%s" % (new_path, action)
path = (action == 'new' and new_path + '.:(format)') or \ "%s.:(format);%s" % (new_path, action)
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
"%s\.:(format);%s" % (member_path, action), **route_options)
"%s.:(format);%s" % (member_path, action), **route_options)
def requirements_for(meth): """Returns a new dict to be used for all route creation as the route options""" opts = options.copy() if method != 'any': opts['conditions'] = {'method':[meth]} return opts
825a0d2a8b34a69b3bec7577699dcc8ab41fd74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12081/825a0d2a8b34a69b3bec7577699dcc8ab41fd74a/base.py
for extrastyle in list(layer.wmsextrastyles):
for extrastyle in [layer.wmsdefaultstyle] + list(layer.wmsextrastyles):
def __init__(self, conf, mapfactory, opsonlineresource): self.conf = conf self.mapfactory = mapfactory if self.conf.has_option('service', 'epsg'): self.crs = CRS('EPSG', self.conf.get('service', 'epsg')) else: ServerConfigurationError('EPSG code not properly configured.')
eea63439c3565666a427f169ae0fe6b82fd99e3d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2302/eea63439c3565666a427f169ae0fe6b82fd99e3d/wms111.py
self._set_servers()
def __init__(self): """Init database""" #Get repos present on machine self.repos = os.listdir("/var/lib/pacman")
192b58976c8291ef317639756a9b5c3e8e6b8cc0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/192b58976c8291ef317639756a9b5c3e8e6b8cc0/pacman.py
return def _set_servers(self): conf_file = open("/etc/pacman.conf", "r").read() self.servers = {} for repo in self.repos: if repo == "foreigners": continue self.servers[repo] = [] self._get_repo_servers(repo, conf_file) continue return def _get_repo_servers(self, repo, conf_file): try: begin = conf_file.index("[%s]" %repo) + len("[%s]" %repo) except ValueError: return end = conf_file.index("[", begin) current_section = conf_file[begin:end].strip() current_lines = current_section.splitlines() current_options = {} for line in current_lines: if not line or line.startswith(" continue opt, value = line.split(" = ") current_options[opt] = value continue for key in current_options.keys(): if key == "Include": servers = open(current_options[key], "r").readlines() for server in servers: server = server.strip() if not server or server.startswith(" continue self.servers[repo].append(server.split(" = ")[1]) continue elif key == "Server": server = current_options[key].split(" = ")[1] self.servers[repo].append(server) continue
def _set_filelist(self, pac, path): """Set installed files list for the given pac""" if not pac.installed: return _("%s is not installed") %pac.name files = open("%s/files" %path).read() begin = files.index("%FILES%") + len("%FILES%") end = files.find("%", begin) - len("%") filelist = files[begin:end].strip() pac.filelist = filelist return
192b58976c8291ef317639756a9b5c3e8e6b8cc0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/192b58976c8291ef317639756a9b5c3e8e6b8cc0/pacman.py
self.show_all()
self.show()
def run(self): self.show_all() self.terminal.fork_command() self.terminal.feed_child("pacman -Sy;exit\n")
0faa1d452f39a60858f6373b8993d931ab096fe5 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/0faa1d452f39a60858f6373b8993d931ab096fe5/dialogs.py
self.vbox.show_all() self.close_button.hide()
def _setup_layout(self):
b5a43c94a23125272d590f68ddde0a01204aa663 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/b5a43c94a23125272d590f68ddde0a01204aa663/dialogs.py
def close(self, widget): self.destroy()
def run(self):
b5a43c94a23125272d590f68ddde0a01204aa663 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/b5a43c94a23125272d590f68ddde0a01204aa663/dialogs.py
self.show_all()
self.show()
def run(self): self.show_all() self.terminal.do_upgrade()
b5a43c94a23125272d590f68ddde0a01204aa663 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/b5a43c94a23125272d590f68ddde0a01204aa663/dialogs.py
if self.database["foreigners"]:
try:
def _setup_pacs_models(self): self.models = {}
b7368ba142cb8d5e19ea6fbb334ce2266458dcd2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/b7368ba142cb8d5e19ea6fbb334ce2266458dcd2/gui.py
self.models["foreigners"] = installed_list(self.database["foreigners"])
if self.database["foreigners"]: self.models["foreigners"] = installed_list(self.database["foreigners"])
def _setup_pacs_models(self): self.models = {}
454b03ed48ac409777149ea4b307465e0a522a78 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/454b03ed48ac409777149ea4b307465e0a522a78/gui.py
elif keyword.count("+"): keys = keyword.split("+")
elif keywords.count("+"): keys = keywords.split("+")
def get_by_keywords(self, keywords): """Return pacs which have keywords as name or in description""" keys = [] #Split keywords by '+' or spaces if keywords.count("+") and keywords.count(" "): keys_1 = keywords.split("+") for key in keys_1: keys = key.split(" ") elif keywords.count(" "): keys = keywords.split() elif keyword.count("+"): keys = keyword.split("+") else: keys = keywords
510cd29a537152d4473b11d04861bf68f6f24a90 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/510cd29a537152d4473b11d04861bf68f6f24a90/pacman.py
self.pages["all"].toggle_renderer.connect("toggled", self.add_install)
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
self.pages["local"]["node"].toggle_renderer.connect("toggled", self.add_install)
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
self.pages["remote"][repo].toggle_renderer.connect("toggled", self.add_install )
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
self.pages["local"][repo].toggle_renderer.connect("toggled", self.add_install)
def _setup_notebook(self): """Setup the notebook""" notebook = self.gld.get_widget("notebook")
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
def _get_selected_page(self, repo_tree):
def _get_selected_page(self, repo_tree, toggle=None):
def _get_selected_page(self, repo_tree): model, tree_iter = repo_tree.get_selection().get_selected() if tree_iter is None: tree_iter = model.get_iter_from_string("0")
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
def add_install(self, widget, data=None):
def add_install(self, widget=None, data=None):
def add_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
return def remove_install(self, widget, data=None):
installed = model.get_value(l_iter, 0) if installed == "yellow" or installed == "green": model.set_value(l_iter, 1, STOCK_GO_UP) else: model.set_value(l_iter, 1, STOCK_ADD) self.queued[name] = (model, l_iter) return def remove_install(self, widget=None, data=None):
def add_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
def remove_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
return def add_remove(self, widget, data=None):
model.set_value(l_iter, 1, "") self.queued.pop(name) return def add_remove(self, widget=None, data=None):
def remove_install(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
def add_remove(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
return def remove_remove(self, widget, data=None):
model.set_value(l_iter, 1, STOCK_REMOVE) self.queued[name] = (model, l_iter) return def remove_remove(self, widget=None, data=None):
def add_remove(self, widget, data=None): repos_tree = self.gld.get_widget("treeview") page = self._get_selected_page(repos_tree)
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
def quit(self, widget, event=None, data=None): main_quit()
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
self.toggle_renderer = CellRendererToggle() self.toggle_renderer.activatable = True
action_renderer = CellRendererPixbuf()
def _setup_columns(self):
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
first_col.pack_start(self.toggle_renderer)
first_col.pack_start(action_renderer)
def _setup_columns(self):
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
first_col.set_attributes(self.toggle_renderer, active=1)
first_col.set_attributes(action_renderer, stock_id=1)
def _setup_columns(self):
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
ListStore.__init__(self, str, bool, str, str, str)
ListStore.__init__(self, str, str, str, str, str)
def __init__(self, packages, repo, node):
8d887e6509e2d71a2f13e91bd822bc9a5fc668ca /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/8d887e6509e2d71a2f13e91bd822bc9a5fc668ca/gui.py
self.gld.get_widget("main_win").show_all()
self._setup_prog_bar()
def __init__(self, fname, icons): """Create the XML Object and connect handlers, then setup notebook""" self.gld = XML(fname, "main_win")
f1686c7d5b259e56020181380ca2f19fe3174339 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/f1686c7d5b259e56020181380ca2f19fe3174339/gui.py
self.repos = os.listdir("/var/lib/pacman") self.repos.remove("local") try: self.repos.remove("wget-log") except ValueError: pass
self.repos = self._get_repos()
def __init__(self): """Init database""" #Get repos present on machine self.repos = os.listdir("/var/lib/pacman")
9f37265ed34a36b81bd35620332047b277d91a06 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2654/9f37265ed34a36b81bd35620332047b277d91a06/pacman.py
print "Found a target: " + str(target) + ", for pattern: " + str(pattern)
def matchTargets(self): rePatternedTargets = {} for pattern in self.patternedTargets.keys(): rePatternedTargets[pattern] = re.compile(pattern)
f147d159c71527bdeb50ccdc541809b1fc6b4bf2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5315/f147d159c71527bdeb50ccdc541809b1fc6b4bf2/make_grapher.py
print "Processing: " + str(path)
def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False):
f147d159c71527bdeb50ccdc541809b1fc6b4bf2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5315/f147d159c71527bdeb50ccdc541809b1fc6b4bf2/make_grapher.py
if allInBetween: for i in range(len(path)): if i < (len(path) - 1): nodes.setdefault(path[i], []) if path[i + 1] not in nodes[path[i]]: nodes[path[i]].append(path[i + 1]) else: if dep not in nodes[path[0]]: nodes[path[0]].append(dep)
print "Matching path: " + str(newpath) if allInBetween: for i in range(len(newpath)): nodes.setdefault(newpath[i], []) if i < (len(newpath) - 1): if newpath[i + 1] not in nodes[newpath[i]]: nodes[newpath[i]].append(newpath[i + 1]) else: if dep not in nodes[path[0]]: nodes[path[0]].append(dep)
def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False):
f147d159c71527bdeb50ccdc541809b1fc6b4bf2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5315/f147d159c71527bdeb50ccdc541809b1fc6b4bf2/make_grapher.py
paths.append(path + [dep])
paths.append(newpath)
def filterNodes(self, seedsIn, seedsOut = None, allInBetween = False):
f147d159c71527bdeb50ccdc541809b1fc6b4bf2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/5315/f147d159c71527bdeb50ccdc541809b1fc6b4bf2/make_grapher.py
def DoxyfileParse(file_contents, dir, data = None):
def DoxyfileParse(file): data = DoxyfileParse_(file,{}) for (k,v) in data.items(): if not v : del data[k] elif k in ("INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES") : continue elif len(v)==1 : data[k] = v[0] return data def DoxyfileParse_(file, data):
def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespace = lex.whitespace.replace("\n", "") lex.escape = "" lineno = lex.lineno token = lex.get_token() key = None last_token = "" key_token = True next_key = False new_data = True def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token while token: if token in ['\n']: if last_token not in ['\\']: key_token = True elif token in ['\\']: pass elif key_token: if key == '@' : key += token else : key = token if token != '@' : key_token = False else: if token == "+=" or (token == "=" and key == "@INCLUDE"): if not data.has_key(key): data[key] = list() elif token == "=": data[key] = list() else: append_data( data, key, new_data, token ) new_data = True if key == '@INCLUDE': inc = os.path.join(dir,data['@INCLUDE'][-1]) if os.path.exists(inc) : DoxyfileParse(open(inc).read(),dir,data) last_token = token token = lex.get_token() if last_token == '\\' and token != '\n': new_data = False append_data( data, key, new_data, '\\' ) # compress lists of len 1 into single strings for (k, v) in data.items(): if len(v) == 0: data.pop(k) # items in the following list will be kept as lists and not converted to strings if k in ["INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES"]: continue if len(v) == 1: data[k] = v[0] return data except: return {}
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
if data is None : data = {}
dir = os.path.dirname(file)
def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespace = lex.whitespace.replace("\n", "") lex.escape = "" lineno = lex.lineno token = lex.get_token() key = None last_token = "" key_token = True next_key = False new_data = True def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token while token: if token in ['\n']: if last_token not in ['\\']: key_token = True elif token in ['\\']: pass elif key_token: if key == '@' : key += token else : key = token if token != '@' : key_token = False else: if token == "+=" or (token == "=" and key == "@INCLUDE"): if not data.has_key(key): data[key] = list() elif token == "=": data[key] = list() else: append_data( data, key, new_data, token ) new_data = True if key == '@INCLUDE': inc = os.path.join(dir,data['@INCLUDE'][-1]) if os.path.exists(inc) : DoxyfileParse(open(inc).read(),dir,data) last_token = token token = lex.get_token() if last_token == '\\' and token != '\n': new_data = False append_data( data, key, new_data, '\\' ) # compress lists of len 1 into single strings for (k, v) in data.items(): if len(v) == 0: data.pop(k) # items in the following list will be kept as lists and not converted to strings if k in ["INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES"]: continue if len(v) == 1: data[k] = v[0] return data except: return {}
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:"
lex = shlex.shlex(instream=open(file), posix=True) lex.wordchars += "*+./-:@~"
def DoxyfileParse(file_contents, dir, data = None): """ Parse a Doxygen source file and return a dictionary of all the values. Values will be strings and lists of strings. """ try: if data is None : data = {} import shlex lex = shlex.shlex(instream = file_contents, posix = True) lex.wordchars += "*+./-:" lex.whitespace = lex.whitespace.replace("\n", "") lex.escape = "" lineno = lex.lineno token = lex.get_token() key = None last_token = "" key_token = True next_key = False new_data = True def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token while token: if token in ['\n']: if last_token not in ['\\']: key_token = True elif token in ['\\']: pass elif key_token: if key == '@' : key += token else : key = token if token != '@' : key_token = False else: if token == "+=" or (token == "=" and key == "@INCLUDE"): if not data.has_key(key): data[key] = list() elif token == "=": data[key] = list() else: append_data( data, key, new_data, token ) new_data = True if key == '@INCLUDE': inc = os.path.join(dir,data['@INCLUDE'][-1]) if os.path.exists(inc) : DoxyfileParse(open(inc).read(),dir,data) last_token = token token = lex.get_token() if last_token == '\\' and token != '\n': new_data = False append_data( data, key, new_data, '\\' ) # compress lists of len 1 into single strings for (k, v) in data.items(): if len(v) == 0: data.pop(k) # items in the following list will be kept as lists and not converted to strings if k in ["INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES"]: continue if len(v) == 1: data[k] = v[0] return data except: return {}
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
if token in ['\n']: if last_token not in ['\\']:
if token=='\n': if last_token!='\\':
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
elif token in ['\\']:
elif token=='\\':
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
if key == '@' : key += token else : key = token if token != '@' : key_token = False
key = token key_token = False
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
if token == "+=" or (token == "=" and key == "@INCLUDE"):
if token=="+=" or (token=="=" and key=="@INCLUDE"):
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
data[key] = list()
data[key] = []
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
data[key] = list()
data[key] = []
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
append_data( data, key, new_data, token )
append_data(data, key, new_data, token)
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
if key == '@INCLUDE':
if key=='@INCLUDE':
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
DoxyfileParse(open(inc).read(),dir,data)
DoxyfileParse_(inc,data)
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
if last_token == '\\' and token != '\n':
if last_token=='\\' and token!='\n':
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
append_data( data, key, new_data, '\\' ) for (k, v) in data.items(): if len(v) == 0: data.pop(k) if k in ["INPUT", "FILE_PATTERNS", "EXCLUDE_PATTERNS", "@INCLUDE", "TAGFILES"]: continue if len(v) == 1: data[k] = v[0]
append_data(data, key, new_data, '\\')
def append_data(data, key, new_data, token): if new_data or len(data[key]) == 0: data[key].append(token) else: data[key][-1] += token
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
dep_add_keys = [
dep_add_keys = (
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx', '*.ipp', '*.i++', '*.inl', '*.h', '*.hh ', '*.hxx', '*.hpp', '*.h++', '*.idl', '*.odl', '*.cs', '*.php', '*.php3', '*.inc', '*.m', '*.mm', '*.py', ] default_exclude_patterns = [ '*~', ] sources = [] basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" ) file_patterns = data.get("FILE_PATTERNS", default_file_patterns) exclude_patterns = data.get("EXCLUDE_PATTERNS", default_exclude_patterns) for i in data.get("INPUT", [ "." ]): input = os.path.normpath(os.path.join(basedir,i)) if os.path.isfile(input): sources.append(input) elif os.path.isdir(input): if recursive : entries = os.walk(input) else : entries = [ (input, [], os.listdir(input)) ] for root, dirs, files in entries: for f in files: filename = os.path.normpath(os.path.join(root, f)) if ( reduce(lambda x, y: x or fnmatch(filename, y), file_patterns, False) and not reduce(lambda x, y: x or fnmatch(filename, y), exclude_patterns, False) ): sources.append(filename) for key in dep_add_keys: if data.has_key(key): elt = data[key] if type(elt) is type ("") : elt = [ elt ] sources.extend([ os.path.normpath(os.path.join(basedir,f)) for f in elt ]) sources = map( lambda path: env.File(path), sources ) return sources
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
]
)
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx', '*.ipp', '*.i++', '*.inl', '*.h', '*.hh ', '*.hxx', '*.hpp', '*.h++', '*.idl', '*.odl', '*.cs', '*.php', '*.php3', '*.inc', '*.m', '*.mm', '*.py', ] default_exclude_patterns = [ '*~', ] sources = [] basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" ) file_patterns = data.get("FILE_PATTERNS", default_file_patterns) exclude_patterns = data.get("EXCLUDE_PATTERNS", default_exclude_patterns) for i in data.get("INPUT", [ "." ]): input = os.path.normpath(os.path.join(basedir,i)) if os.path.isfile(input): sources.append(input) elif os.path.isdir(input): if recursive : entries = os.walk(input) else : entries = [ (input, [], os.listdir(input)) ] for root, dirs, files in entries: for f in files: filename = os.path.normpath(os.path.join(root, f)) if ( reduce(lambda x, y: x or fnmatch(filename, y), file_patterns, False) and not reduce(lambda x, y: x or fnmatch(filename, y), exclude_patterns, False) ): sources.append(filename) for key in dep_add_keys: if data.has_key(key): elt = data[key] if type(elt) is type ("") : elt = [ elt ] sources.extend([ os.path.normpath(os.path.join(basedir,f)) for f in elt ]) sources = map( lambda path: env.File(path), sources ) return sources
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
default_file_patterns = [
default_file_patterns = (
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx', '*.ipp', '*.i++', '*.inl', '*.h', '*.hh ', '*.hxx', '*.hpp', '*.h++', '*.idl', '*.odl', '*.cs', '*.php', '*.php3', '*.inc', '*.m', '*.mm', '*.py', ] default_exclude_patterns = [ '*~', ] sources = [] basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" ) file_patterns = data.get("FILE_PATTERNS", default_file_patterns) exclude_patterns = data.get("EXCLUDE_PATTERNS", default_exclude_patterns) for i in data.get("INPUT", [ "." ]): input = os.path.normpath(os.path.join(basedir,i)) if os.path.isfile(input): sources.append(input) elif os.path.isdir(input): if recursive : entries = os.walk(input) else : entries = [ (input, [], os.listdir(input)) ] for root, dirs, files in entries: for f in files: filename = os.path.normpath(os.path.join(root, f)) if ( reduce(lambda x, y: x or fnmatch(filename, y), file_patterns, False) and not reduce(lambda x, y: x or fnmatch(filename, y), exclude_patterns, False) ): sources.append(filename) for key in dep_add_keys: if data.has_key(key): elt = data[key] if type(elt) is type ("") : elt = [ elt ] sources.extend([ os.path.normpath(os.path.join(basedir,f)) for f in elt ]) sources = map( lambda path: env.File(path), sources ) return sources
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
] default_exclude_patterns = [
) default_exclude_patterns = (
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx', '*.ipp', '*.i++', '*.inl', '*.h', '*.hh ', '*.hxx', '*.hpp', '*.h++', '*.idl', '*.odl', '*.cs', '*.php', '*.php3', '*.inc', '*.m', '*.mm', '*.py', ] default_exclude_patterns = [ '*~', ] sources = [] basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" ) file_patterns = data.get("FILE_PATTERNS", default_file_patterns) exclude_patterns = data.get("EXCLUDE_PATTERNS", default_exclude_patterns) for i in data.get("INPUT", [ "." ]): input = os.path.normpath(os.path.join(basedir,i)) if os.path.isfile(input): sources.append(input) elif os.path.isdir(input): if recursive : entries = os.walk(input) else : entries = [ (input, [], os.listdir(input)) ] for root, dirs, files in entries: for f in files: filename = os.path.normpath(os.path.join(root, f)) if ( reduce(lambda x, y: x or fnmatch(filename, y), file_patterns, False) and not reduce(lambda x, y: x or fnmatch(filename, y), exclude_patterns, False) ): sources.append(filename) for key in dep_add_keys: if data.has_key(key): elt = data[key] if type(elt) is type ("") : elt = [ elt ] sources.extend([ os.path.normpath(os.path.join(basedir,f)) for f in elt ]) sources = map( lambda path: env.File(path), sources ) return sources
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
]
)
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx', '*.ipp', '*.i++', '*.inl', '*.h', '*.hh ', '*.hxx', '*.hpp', '*.h++', '*.idl', '*.odl', '*.cs', '*.php', '*.php3', '*.inc', '*.m', '*.mm', '*.py', ] default_exclude_patterns = [ '*~', ] sources = [] basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" ) file_patterns = data.get("FILE_PATTERNS", default_file_patterns) exclude_patterns = data.get("EXCLUDE_PATTERNS", default_exclude_patterns) for i in data.get("INPUT", [ "." ]): input = os.path.normpath(os.path.join(basedir,i)) if os.path.isfile(input): sources.append(input) elif os.path.isdir(input): if recursive : entries = os.walk(input) else : entries = [ (input, [], os.listdir(input)) ] for root, dirs, files in entries: for f in files: filename = os.path.normpath(os.path.join(root, f)) if ( reduce(lambda x, y: x or fnmatch(filename, y), file_patterns, False) and not reduce(lambda x, y: x or fnmatch(filename, y), exclude_patterns, False) ): sources.append(filename) for key in dep_add_keys: if data.has_key(key): elt = data[key] if type(elt) is type ("") : elt = [ elt ] sources.extend([ os.path.normpath(os.path.join(basedir,f)) for f in elt ]) sources = map( lambda path: env.File(path), sources ) return sources
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" )
basedir = node.dir.abspath data = DoxyfileParse(node.abspath) recursive = data.get("RECURSIVE", "NO").upper()=="YES"
def DoxySourceScan(node, env, path): """ Doxygen Doxyfile source scanner. This should scan the Doxygen file and add any files used to generate docs to the list of source files. """ dep_add_keys = [ '@INCLUDE', 'HTML_HEADER', 'HTML_FOOTER', 'TAGFILES' ] default_file_patterns = [ '*.c', '*.cc', '*.cxx', '*.cpp', '*.c++', '*.java', '*.ii', '*.ixx', '*.ipp', '*.i++', '*.inl', '*.h', '*.hh ', '*.hxx', '*.hpp', '*.h++', '*.idl', '*.odl', '*.cs', '*.php', '*.php3', '*.inc', '*.m', '*.mm', '*.py', ] default_exclude_patterns = [ '*~', ] sources = [] basedir = str(node.dir) data = DoxyfileParse(node.get_contents(), basedir) recursive = ( data.get("RECURSIVE", "NO") == "YES" ) file_patterns = data.get("FILE_PATTERNS", default_file_patterns) exclude_patterns = data.get("EXCLUDE_PATTERNS", default_exclude_patterns) for i in data.get("INPUT", [ "." ]): input = os.path.normpath(os.path.join(basedir,i)) if os.path.isfile(input): sources.append(input) elif os.path.isdir(input): if recursive : entries = os.walk(input) else : entries = [ (input, [], os.listdir(input)) ] for root, dirs, files in entries: for f in files: filename = os.path.normpath(os.path.join(root, f)) if ( reduce(lambda x, y: x or fnmatch(filename, y), file_patterns, False) and not reduce(lambda x, y: x or fnmatch(filename, y), exclude_patterns, False) ): sources.append(filename) for key in dep_add_keys: if data.has_key(key): elt = data[key] if type(elt) is type ("") : elt = [ elt ] sources.extend([ os.path.normpath(os.path.join(basedir,f)) for f in elt ]) sources = map( lambda path: env.File(path), sources ) return sources
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
"HTML": ("YES", "html"), "LATEX": ("YES", "latex"), "RTF": ("NO", "rtf"), "MAN": ("YES", "man"), "XML": ("NO", "xml"),
"HTML" : ("YES", "html"), "LATEX" : ("YES", "latex"), "RTF" : ("NO", "rtf"), "MAN" : ("YES", "man"), "XML" : ("NO", "xml"),
def DoxyEmitter(source, target, env): """Doxygen Doxyfile emitter""" # possible output formats and their default values and output locations output_formats = { "HTML": ("YES", "html"), "LATEX": ("YES", "latex"), "RTF": ("NO", "rtf"), "MAN": ("YES", "man"), "XML": ("NO", "xml"), } data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) targets = [] out_dir = data.get("OUTPUT_DIRECTORY", ".") # add our output locations for (k, v) in output_formats.items(): if data.get("GENERATE_" + k, v[0]) == "YES": # Grmpf ... need to use a File object here. The problem is, that # Dir.scan() is implemented to just return the directory entries # and does *not* invoke the source-file scanners .. ARGH !! dir = env.Dir( os.path.join(str(source[0].dir), out_dir, data.get(k + "_OUTPUT", v[1])) ) node = env.File( os.path.join(str(dir), ".stamp" ) ) env.Clean(node, dir) targets.append( node ) if data.has_key("GENERATE_TAGFILE"): targets.append(env.File( os.path.join(str(source[0].dir), data["GENERATE_TAGFILE"]) )) # don't clobber targets for node in targets: env.Precious(node) return (targets, source)
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
data = DoxyfileParse(source[0].get_contents(), str(source[0].dir))
data = DoxyfileParse(source[0].abspath)
def DoxyEmitter(source, target, env): """Doxygen Doxyfile emitter""" # possible output formats and their default values and output locations output_formats = { "HTML": ("YES", "html"), "LATEX": ("YES", "latex"), "RTF": ("NO", "rtf"), "MAN": ("YES", "man"), "XML": ("NO", "xml"), } data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) targets = [] out_dir = data.get("OUTPUT_DIRECTORY", ".") # add our output locations for (k, v) in output_formats.items(): if data.get("GENERATE_" + k, v[0]) == "YES": # Grmpf ... need to use a File object here. The problem is, that # Dir.scan() is implemented to just return the directory entries # and does *not* invoke the source-file scanners .. ARGH !! dir = env.Dir( os.path.join(str(source[0].dir), out_dir, data.get(k + "_OUTPUT", v[1])) ) node = env.File( os.path.join(str(dir), ".stamp" ) ) env.Clean(node, dir) targets.append( node ) if data.has_key("GENERATE_TAGFILE"): targets.append(env.File( os.path.join(str(source[0].dir), data["GENERATE_TAGFILE"]) )) # don't clobber targets for node in targets: env.Precious(node) return (targets, source)
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
for (k, v) in output_formats.items(): if data.get("GENERATE_" + k, v[0]) == "YES":
for (k, v) in output_formats.iteritems(): if data.get("GENERATE_" + k, v[0]).upper() == "YES":
def DoxyEmitter(source, target, env): """Doxygen Doxyfile emitter""" # possible output formats and their default values and output locations output_formats = { "HTML": ("YES", "html"), "LATEX": ("YES", "latex"), "RTF": ("NO", "rtf"), "MAN": ("YES", "man"), "XML": ("NO", "xml"), } data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) targets = [] out_dir = data.get("OUTPUT_DIRECTORY", ".") # add our output locations for (k, v) in output_formats.items(): if data.get("GENERATE_" + k, v[0]) == "YES": # Grmpf ... need to use a File object here. The problem is, that # Dir.scan() is implemented to just return the directory entries # and does *not* invoke the source-file scanners .. ARGH !! dir = env.Dir( os.path.join(str(source[0].dir), out_dir, data.get(k + "_OUTPUT", v[1])) ) node = env.File( os.path.join(str(dir), ".stamp" ) ) env.Clean(node, dir) targets.append( node ) if data.has_key("GENERATE_TAGFILE"): targets.append(env.File( os.path.join(str(source[0].dir), data["GENERATE_TAGFILE"]) )) # don't clobber targets for node in targets: env.Precious(node) return (targets, source)
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
data = DoxyfileParse(node.sources[0].get_contents(), str(node.sources[0].dir)) if data.get("GENERATE_HTML",'YES') != 'YES' : return None return os.path.normpath(os.path.join( str(node.sources[0].dir),
data = DoxyfileParse(node.sources[0].abspath) if data.get("GENERATE_HTML",'YES').upper() != 'YES' : return None return os.path.normpath(os.path.join( node.sources[0].abspath,
def doxyNodeHtmlDir(node): if not node.sources : return None data = DoxyfileParse(node.sources[0].get_contents(), str(node.sources[0].dir)) if data.get("GENERATE_HTML",'YES') != 'YES' : return None return os.path.normpath(os.path.join( str(node.sources[0].dir), data.get("OUTPUT_DIRECTORY","."), data.get("HTML_OUTPUT","html") ))
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
data = DoxyfileParse(source[0].get_contents(), str(source[0].dir))
data = DoxyfileParse(source[0].abspath)
def DoxyGenerator(source, target, env, for_signature): data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) actions = [ env.Action("cd ${SOURCE.dir} && ${DOXYGEN} ${SOURCE.file}") ] # This will add automatic 'installdox' calls. # # For every referenced tagfile, the generator first checks for the # existence of a construction variable '<name>_DOXY_URL' where # '<name>' is the uppercased name of the tagfile sans extension # (e.g. 'Utils.tag' -> 'UTILS_DOXY_URL'). If this variable exists, # it must contain the url or path to the installed documentation # corresponding to the tag file. # # Is the variable is not found and if a referenced tag file is a # target within this same build, the generator will parse the # 'Doxyfile' from which the tag file is built. It will # automatically create the html directory from the information in # that 'Doxyfile'. # # If for any referenced tagfile no url can be found, 'installdox' # will *not* be called and a warning about the missing url is # generated. if data.get('GENERATE_HTML','YES') == "YES": output_dir = os.path.normpath(os.path.join( str(source[0].dir), data.get("OUTPUT_DIRECTORY","."), data.get("HTML_OUTPUT","html") )) args = [] for tagfile in data.get('TAGFILES',[]): url = env.get(os.path.splitext(os.path.basename(tagfile))[0].upper()+"_DOXY_URL", None) if not url: url = doxyNodeHtmlDir( env.File(os.path.normpath(os.path.join( str(source[0].dir), tagfile )))) if url : url = relpath(output_dir, url) if not url: print "WARNING:",str(node.sources[0]),": missing tagfile url for",tagfile args = None if args is not None and url: args.append("-l %s@%s" % ( os.path.basename(tagfile), url )) if args: actions.append(env.Action('cd %s && ./installdox %s' % (output_dir, " ".join(args)))) actions.append(env.Action([ "touch $TARGETS" ])) return actions
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
if data.get('GENERATE_HTML','YES') == "YES": output_dir = os.path.normpath(os.path.join( str(source[0].dir),
if data.get('GENERATE_HTML','YES').upper() == "YES": output_dir = os.path.normpath(os.path.join( source[0].dir.abspath,
def DoxyGenerator(source, target, env, for_signature): data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) actions = [ env.Action("cd ${SOURCE.dir} && ${DOXYGEN} ${SOURCE.file}") ] # This will add automatic 'installdox' calls. # # For every referenced tagfile, the generator first checks for the # existence of a construction variable '<name>_DOXY_URL' where # '<name>' is the uppercased name of the tagfile sans extension # (e.g. 'Utils.tag' -> 'UTILS_DOXY_URL'). If this variable exists, # it must contain the url or path to the installed documentation # corresponding to the tag file. # # Is the variable is not found and if a referenced tag file is a # target within this same build, the generator will parse the # 'Doxyfile' from which the tag file is built. It will # automatically create the html directory from the information in # that 'Doxyfile'. # # If for any referenced tagfile no url can be found, 'installdox' # will *not* be called and a warning about the missing url is # generated. if data.get('GENERATE_HTML','YES') == "YES": output_dir = os.path.normpath(os.path.join( str(source[0].dir), data.get("OUTPUT_DIRECTORY","."), data.get("HTML_OUTPUT","html") )) args = [] for tagfile in data.get('TAGFILES',[]): url = env.get(os.path.splitext(os.path.basename(tagfile))[0].upper()+"_DOXY_URL", None) if not url: url = doxyNodeHtmlDir( env.File(os.path.normpath(os.path.join( str(source[0].dir), tagfile )))) if url : url = relpath(output_dir, url) if not url: print "WARNING:",str(node.sources[0]),": missing tagfile url for",tagfile args = None if args is not None and url: args.append("-l %s@%s" % ( os.path.basename(tagfile), url )) if args: actions.append(env.Action('cd %s && ./installdox %s' % (output_dir, " ".join(args)))) actions.append(env.Action([ "touch $TARGETS" ])) return actions
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
env.File(os.path.normpath(os.path.join( str(source[0].dir), tagfile ))))
env.File(os.path.normpath(os.path.join(str(source[0].dir), tagfile))))
def DoxyGenerator(source, target, env, for_signature): data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) actions = [ env.Action("cd ${SOURCE.dir} && ${DOXYGEN} ${SOURCE.file}") ] # This will add automatic 'installdox' calls. # # For every referenced tagfile, the generator first checks for the # existence of a construction variable '<name>_DOXY_URL' where # '<name>' is the uppercased name of the tagfile sans extension # (e.g. 'Utils.tag' -> 'UTILS_DOXY_URL'). If this variable exists, # it must contain the url or path to the installed documentation # corresponding to the tag file. # # Is the variable is not found and if a referenced tag file is a # target within this same build, the generator will parse the # 'Doxyfile' from which the tag file is built. It will # automatically create the html directory from the information in # that 'Doxyfile'. # # If for any referenced tagfile no url can be found, 'installdox' # will *not* be called and a warning about the missing url is # generated. if data.get('GENERATE_HTML','YES') == "YES": output_dir = os.path.normpath(os.path.join( str(source[0].dir), data.get("OUTPUT_DIRECTORY","."), data.get("HTML_OUTPUT","html") )) args = [] for tagfile in data.get('TAGFILES',[]): url = env.get(os.path.splitext(os.path.basename(tagfile))[0].upper()+"_DOXY_URL", None) if not url: url = doxyNodeHtmlDir( env.File(os.path.normpath(os.path.join( str(source[0].dir), tagfile )))) if url : url = relpath(output_dir, url) if not url: print "WARNING:",str(node.sources[0]),": missing tagfile url for",tagfile args = None if args is not None and url: args.append("-l %s@%s" % ( os.path.basename(tagfile), url )) if args: actions.append(env.Action('cd %s && ./installdox %s' % (output_dir, " ".join(args)))) actions.append(env.Action([ "touch $TARGETS" ])) return actions
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
print "WARNING:",str(node.sources[0]),": missing tagfile url for",tagfile
print "WARNING:",source[0].abspath, ": missing tagfile url for", tagfile
def DoxyGenerator(source, target, env, for_signature): data = DoxyfileParse(source[0].get_contents(), str(source[0].dir)) actions = [ env.Action("cd ${SOURCE.dir} && ${DOXYGEN} ${SOURCE.file}") ] # This will add automatic 'installdox' calls. # # For every referenced tagfile, the generator first checks for the # existence of a construction variable '<name>_DOXY_URL' where # '<name>' is the uppercased name of the tagfile sans extension # (e.g. 'Utils.tag' -> 'UTILS_DOXY_URL'). If this variable exists, # it must contain the url or path to the installed documentation # corresponding to the tag file. # # Is the variable is not found and if a referenced tag file is a # target within this same build, the generator will parse the # 'Doxyfile' from which the tag file is built. It will # automatically create the html directory from the information in # that 'Doxyfile'. # # If for any referenced tagfile no url can be found, 'installdox' # will *not* be called and a warning about the missing url is # generated. if data.get('GENERATE_HTML','YES') == "YES": output_dir = os.path.normpath(os.path.join( str(source[0].dir), data.get("OUTPUT_DIRECTORY","."), data.get("HTML_OUTPUT","html") )) args = [] for tagfile in data.get('TAGFILES',[]): url = env.get(os.path.splitext(os.path.basename(tagfile))[0].upper()+"_DOXY_URL", None) if not url: url = doxyNodeHtmlDir( env.File(os.path.normpath(os.path.join( str(source[0].dir), tagfile )))) if url : url = relpath(output_dir, url) if not url: print "WARNING:",str(node.sources[0]),": missing tagfile url for",tagfile args = None if args is not None and url: args.append("-l %s@%s" % ( os.path.basename(tagfile), url )) if args: actions.append(env.Action('cd %s && ./installdox %s' % (output_dir, " ".join(args)))) actions.append(env.Action([ "touch $TARGETS" ])) return actions
1be18a2af8b3880d492f80349c68c4d9cf1ae16a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/1be18a2af8b3880d492f80349c68c4d9cf1ae16a/Doxygen.py
return env.Command(target, testRunner, [ './$SOURCE $BOOSTTESTARGS | tee $TARGET' ])
return env.Command([ target, stamp ], testRunner, [ '( $SOURCE $BOOSTTESTARGS 2>&1 && touch ${TARGETS[1]} ) | tee ${TARGETS[0]}' ])
def BoostUnitTests(env, target, source, test_source=None, LIBS = [], DEPENDS = [], **kw): path, name = os.path.split(target) if test_source: if type(test_source) is not type([]): test_source = [ test_source ] else: test_source = [] testEnv = env.Copy(**kw) testEnv.Append(LIBS = '$BOOSTTESTLIB') testEnv.Append(LIBS = LIBS) sources = [] if source: sources = sources + env.Object(source) sources = sources + test_source binName = os.path.join(path,'.' + os.path.splitext(name)[0]+'.bin') testRunner = testEnv.Program(binName, sources) if DEPENDS: env.Depends(testRunner, DEPENDS) return env.Command(target, testRunner, [ './$SOURCE $BOOSTTESTARGS | tee $TARGET' ])
c192012710b7ee6e5e88af29b119465f58e039a5 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/c192012710b7ee6e5e88af29b119465f58e039a5/BoostUnitTests.py
def FinalizeBoost(env): env.Tool('BoostUnitTests', [os.path.split(__file__)[0]]) if env['BOOST_TOOLSET']: runtime = "" if not env['final'] : runtime += "gd" if env['STLPORT_LIB'] : runtime += "p" if runtime: runtime = "-" + runtime env['BOOST_VARIANT'] = "-" + env['BOOST_TOOLSET'] + runtime env['BOOSTTESTLIB'] = 'libboost_unit_test_framework' + env['BOOST_VARIANT'] env.Append(LIBPATH = [ '$BOOST_LIBDIR' ], CPPPATH = [ '$BOOST_INCLUDES' ])
3ed7e36cc3d0064c7c2c513a5d834826eb01807f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/3ed7e36cc3d0064c7c2c513a5d834826eb01807f/SatSCons.py
'-pedantic', '-ansi' ],
'-pedantic', '-ansi' ],
def MakeEnvironment(): global opts, finalizers InitOpts() env = SCons.Environment.Environment(options=opts) if SCons.Script.SConscript.Arguments.get('final'): env['final'] = 1 env.Help(opts.GenerateHelpText(env)) #conf = env.Configure() #env = conf.env if os.environ.has_key('SSH_AUTH_SOCK'): env.Append( ENV = { 'SSH_AUTH_SOCK': os.environ['SSH_AUTH_SOCK'] } ) for finalizer in finalizers: finalizer(env) env.Append(CXXFLAGS = [ '-Wall', '-Woverloaded-virtual', '-Wno-long-long', '-pedantic', '-ansi' ], LOCALLIBDIR = [ '#' ], LIBPATH = [ '$LOCALLIBDIR' ]) if env['final']: env.Append(CXXFLAGS = [ '-O3', '-g' ], LINKFLAGS = [ '-g' ]) else: env.Append(CXXFLAGS = [ '-O0', '-g', '-fno-inline' ], LINKFLAGS = [ '-g' ]) #return conf.Finish() return env
3ed7e36cc3d0064c7c2c513a5d834826eb01807f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/3ed7e36cc3d0064c7c2c513a5d834826eb01807f/SatSCons.py
env.Append(CXXFLAGS = [ '-O3', '-g' ], LINKFLAGS = [ '-g' ])
env.Append(CXXFLAGS = [ '-O3' ], CPPDEFINES = [ 'NDEBUG' ])
def MakeEnvironment(): global opts, finalizers InitOpts() env = SCons.Environment.Environment(options=opts) if SCons.Script.SConscript.Arguments.get('final'): env['final'] = 1 env.Help(opts.GenerateHelpText(env)) #conf = env.Configure() #env = conf.env if os.environ.has_key('SSH_AUTH_SOCK'): env.Append( ENV = { 'SSH_AUTH_SOCK': os.environ['SSH_AUTH_SOCK'] } ) for finalizer in finalizers: finalizer(env) env.Append(CXXFLAGS = [ '-Wall', '-Woverloaded-virtual', '-Wno-long-long', '-pedantic', '-ansi' ], LOCALLIBDIR = [ '#' ], LIBPATH = [ '$LOCALLIBDIR' ]) if env['final']: env.Append(CXXFLAGS = [ '-O3', '-g' ], LINKFLAGS = [ '-g' ]) else: env.Append(CXXFLAGS = [ '-O0', '-g', '-fno-inline' ], LINKFLAGS = [ '-g' ]) #return conf.Finish() return env
3ed7e36cc3d0064c7c2c513a5d834826eb01807f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/3ed7e36cc3d0064c7c2c513a5d834826eb01807f/SatSCons.py
testEnv.Append(LIBS = '$BOOSTTESTLIB') testEnv.Append(LIBS = LIBS)
testEnv.Prepend(LIBS = '$BOOSTTESTLIB') testEnv.Prepend(LIBS = LIBS)
def BoostUnitTests(env, target, source, test_source=None, LIBS = [], DEPENDS = [], **kw): path, name = os.path.split(target) if test_source: if type(test_source) is not type([]): test_source = [ test_source ] else: test_source = [] testEnv = env.Copy(**kw) testEnv.Append(LIBS = '$BOOSTTESTLIB') testEnv.Append(LIBS = LIBS) sources = [] if source: sources = sources + env.Object(source) sources = sources + test_source binName = os.path.join(path,'.' + os.path.splitext(name)[0]+'.bin') testRunner = testEnv.Program(binName, sources) stamp = os.path.join(path,'.' + os.path.splitext(name)[0]+'.stamp') if DEPENDS: env.Depends(testRunner, DEPENDS) return env.Command([ target, stamp ], testRunner, [ '( $SOURCE $BOOSTTESTARGS 2>&1 && touch ${TARGETS[1]} ) | tee ${TARGETS[0]}' ])
a83afd21a4b14abe0604a1c5f344cb33f92ee9fd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/a83afd21a4b14abe0604a1c5f344cb33f92ee9fd/BoostUnitTests.py
if not env['final'] : runtime += "gd"
if env['final'] : runtime += env.get('BOOST_RUNTIME','') else : runtime += env.get('BOOST_DEBUG_RUNTIME','gd')
def FinalizeBoost(env): env.Tool('BoostUnitTests', [os.path.split(__file__)[0]]) if env['BOOST_TOOLSET']: runtime = "" if not env['final'] : runtime += "gd" if env['STLPORT_LIB'] : runtime += "p" if runtime: runtime = "-" + runtime env['BOOST_VARIANT'] = "-" + env['BOOST_TOOLSET'] + runtime env['BOOSTTESTLIB'] = 'libboost_unit_test_framework' + env['BOOST_VARIANT'] env.Append(LIBPATH = [ '$BOOST_LIBDIR' ], CPPPATH = [ '$BOOST_INCLUDES' ])
10a735f2839053552c177ebba0de1980487b24f8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/10a735f2839053552c177ebba0de1980487b24f8/SatSCons.py
env['STLPORT_DEBUGLIB'] = ''
def FinalizeSTLPort(env): env['STLPORT_DEBUGLIB'] = '' if env['STLPORT_LIB']: env['STLPORT_DEBUGLIB'] = env['STLPORT_LIB'] + '_stldebug' env.Append(LIBPATH = [ '$STLPORT_LIBDIR' ], CPPPATH = [ '$STLPORT_INCLUDES' ]) if env['final']: env.Append(LIBS = [ '$STLPORT_LIB' ]) else: env.Append(LIBS = [ '$STLPORT_DEBUGLIB' ], CPPDEFINES = [ '_STLP_DEBUG' ])
10a735f2839053552c177ebba0de1980487b24f8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/10a735f2839053552c177ebba0de1980487b24f8/SatSCons.py
env['STLPORT_DEBUGLIB'] = env['STLPORT_LIB'] + '_stldebug'
if not env['STLPORT_DEBUGLIB']: env['STLPORT_DEBUGLIB'] = env['STLPORT_LIB'] + '_stldebug'
def FinalizeSTLPort(env): env['STLPORT_DEBUGLIB'] = '' if env['STLPORT_LIB']: env['STLPORT_DEBUGLIB'] = env['STLPORT_LIB'] + '_stldebug' env.Append(LIBPATH = [ '$STLPORT_LIBDIR' ], CPPPATH = [ '$STLPORT_INCLUDES' ]) if env['final']: env.Append(LIBS = [ '$STLPORT_LIB' ]) else: env.Append(LIBS = [ '$STLPORT_DEBUGLIB' ], CPPDEFINES = [ '_STLP_DEBUG' ])
10a735f2839053552c177ebba0de1980487b24f8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/10a735f2839053552c177ebba0de1980487b24f8/SatSCons.py
env.Append(CXXFLAGS = [ '-Wall', '-Woverloaded-virtual', '-Wno-long-long', '-pedantic', '-ansi' ],
env.Append(CXXFLAGS = [ '-Wall', '-Woverloaded-virtual', '-Wno-long-long' ],
def MakeEnvironment(): global opts, finalizers InitOpts() env = SCons.Environment.Environment(options=opts) if SCons.Script.SConscript.Arguments.get('final'): env['final'] = 1 env.Help(opts.GenerateHelpText(env)) #conf = env.Configure() #env = conf.env if os.environ.has_key('SSH_AUTH_SOCK'): env.Append( ENV = { 'SSH_AUTH_SOCK': os.environ['SSH_AUTH_SOCK'] } ) for finalizer in finalizers: finalizer(env) env.Append(CXXFLAGS = [ '-Wall', '-Woverloaded-virtual', '-Wno-long-long', '-pedantic', '-ansi' ], LOCALLIBDIR = [ '#' ], LIBPATH = [ '$LOCALLIBDIR' ]) if env['final']: env.Append(CXXFLAGS = [ '-O3' ], CPPDEFINES = [ 'NDEBUG' ]) else: env.Append(CXXFLAGS = [ '-O0', '-g', '-fno-inline' ], LINKFLAGS = [ '-g' ]) #return conf.Finish() return env
10a735f2839053552c177ebba0de1980487b24f8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/10a735f2839053552c177ebba0de1980487b24f8/SatSCons.py
LINKFLAGS = [ '-g' ])
LINKFLAGS = [ '-g' ]) env.Append(CPPDEFINES = [ '$EXTRA_DEFINES' ], LIBS = [ '$EXTRA_LIBS' ])
def MakeEnvironment(): global opts, finalizers InitOpts() env = SCons.Environment.Environment(options=opts) if SCons.Script.SConscript.Arguments.get('final'): env['final'] = 1 env.Help(opts.GenerateHelpText(env)) #conf = env.Configure() #env = conf.env if os.environ.has_key('SSH_AUTH_SOCK'): env.Append( ENV = { 'SSH_AUTH_SOCK': os.environ['SSH_AUTH_SOCK'] } ) for finalizer in finalizers: finalizer(env) env.Append(CXXFLAGS = [ '-Wall', '-Woverloaded-virtual', '-Wno-long-long', '-pedantic', '-ansi' ], LOCALLIBDIR = [ '#' ], LIBPATH = [ '$LOCALLIBDIR' ]) if env['final']: env.Append(CXXFLAGS = [ '-O3' ], CPPDEFINES = [ 'NDEBUG' ]) else: env.Append(CXXFLAGS = [ '-O0', '-g', '-fno-inline' ], LINKFLAGS = [ '-g' ]) #return conf.Finish() return env
10a735f2839053552c177ebba0de1980487b24f8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/10a735f2839053552c177ebba0de1980487b24f8/SatSCons.py
[ '( $SOURCE $BOOSTTESTARGS 2>&1 && touch ${TARGETS[1]} ) | tee ${TARGETS[0]}' ])
[ '( $SOURCE $BOOSTTESTARGS 2>&1 && touch ${TARGETS[1]} ) | tee ${TARGETS[0]}; exit $$PIPESTATUS[0]' ])
def BoostUnitTests(env, target, source, test_source=None, LIBS = [], DEPENDS = [], **kw): path, name = os.path.split(target) if test_source: if type(test_source) is not type([]): test_source = [ test_source ] else: test_source = [] testEnv = env.Copy(**kw) testEnv.Prepend(LIBS = '$BOOSTTESTLIB') testEnv.Prepend(LIBS = LIBS) sources = [] if source: sources = sources + env.Object(source) sources = sources + test_source binName = os.path.join(path,'.' + os.path.splitext(name)[0]+'.bin') testRunner = testEnv.Program(binName, sources) stamp = os.path.join(path,'.' + os.path.splitext(name)[0]+'.stamp') if DEPENDS: env.Depends(testRunner, DEPENDS) return env.Command([ target, stamp ], testRunner, [ '( $SOURCE $BOOSTTESTARGS 2>&1 && touch ${TARGETS[1]} ) | tee ${TARGETS[0]}' ])
2adcf23c3aad3966b0dd3ea1c5de34c50184e7eb /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2548/2adcf23c3aad3966b0dd3ea1c5de34c50184e7eb/BoostUnitTests.py
if callback: _sqlitecache.log_handler_add(callback.log)
self.callback = callback
def __init__(self, storedir, repoid, callback=None):
d56127a5ce286fc95e5f57bb574fefc23a7db216 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2779/d56127a5ce286fc95e5f57bb574fefc23a7db216/sqlitecache.py
return self.open_database(_sqlitecache.update_primary(location, checksum))
return self.open_database(_sqlitecache.update_primary(location, checksum, self.callback))
def getPrimary(self, location, checksum): """Load primary.xml.gz from an sqlite cache and update it if required""" return self.open_database(_sqlitecache.update_primary(location, checksum))
d56127a5ce286fc95e5f57bb574fefc23a7db216 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2779/d56127a5ce286fc95e5f57bb574fefc23a7db216/sqlitecache.py
return self.open_database(_sqlitecache.update_filelist(location, checksum))
return self.open_database(_sqlitecache.update_filelist(location, checksum, self.callback))
def getFilelists(self, location, checksum): """Load filelist.xml.gz from an sqlite cache and update it if required""" return self.open_database(_sqlitecache.update_filelist(location, checksum))
d56127a5ce286fc95e5f57bb574fefc23a7db216 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2779/d56127a5ce286fc95e5f57bb574fefc23a7db216/sqlitecache.py
host, port = urllib.splitnport(host)
newhost, port = urllib.splitnport(host) if port is not None: host = newhost
def url_fix_host (urlparts): """ Unquote and fix hostname. Returns is_idn. """ urlparts[1], is_idn = idna_encode(urllib.unquote(urlparts[1]).lower()) # a leading backslash in path causes urlsplit() to add the # path components up to the first slash to host # try to find this case... i = urlparts[1].find("\\") if i != -1: # ...and fix it by prepending the misplaced components to the path comps = urlparts[1][i:] # note: still has leading backslash if not urlparts[2] or urlparts[2] == '/': urlparts[2] = comps else: urlparts[2] = "%s%s" % (comps, urllib.unquote(urlparts[2])) urlparts[1] = urlparts[1][:i] else: # a leading ? in path causes urlsplit() to add the query to the # host name i = urlparts[1].find("?") if i != -1: urlparts[1], urlparts[3] = urlparts[1].split('?', 1) # path urlparts[2] = urllib.unquote(urlparts[2]) if urlparts[1]: userpass, host = urllib.splituser(urlparts[1]) if userpass: # append AT for easy concatenation userpass += "@" else: userpass = "" host, port = urllib.splitnport(host) # remove trailing dot if host.endswith("."): host = host[:-1] # remove a default (or invalid) port if port in [-1, None, default_ports.get(urlparts[0])]: urlparts[1] = userpass+host else: urlparts[1] = "%s%s:%d" % (userpass, host, port)
aa523745b34373f1092ec5c52339619ffc90bf51 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3948/aa523745b34373f1092ec5c52339619ffc90bf51/url.py