2017-07-30 20:11:29 +02:00
|
|
|
#! /usr/bin/env python
|
|
|
|
# encoding: utf-8
|
|
|
|
# Thomas Nagy, 2017 (ita)
|
|
|
|
|
|
|
|
"""
|
|
|
|
A system for fast partial rebuilds
|
|
|
|
|
|
|
|
Creating a large amount of task objects up front can take some time.
|
|
|
|
By making a few assumptions, it is possible to avoid posting creating
|
|
|
|
task objects for targets that are already up-to-date
|
|
|
|
|
|
|
|
On a silly benchmark the gain observed for 20000 tasks can be 5s->1s for
|
|
|
|
a single file change.
|
|
|
|
|
|
|
|
Assuptions:
|
|
|
|
* Mostly for C/C++/Fortran targets with link tasks (object-only targets are not handled)
|
2017-08-02 01:23:32 +02:00
|
|
|
* For full project builds: no --targets and no pruning from subfolders
|
2017-07-30 20:11:29 +02:00
|
|
|
* The installation phase is ignored
|
|
|
|
* `use=` dependencies are fully specified up front even across build groups
|
2017-08-02 01:23:32 +02:00
|
|
|
* Task generator source files are not obtained from globs
|
2017-07-30 20:11:29 +02:00
|
|
|
"""
|
|
|
|
|
|
|
|
import os
|
2017-08-02 20:05:00 +02:00
|
|
|
from waflib import Build, Context, Errors, Logs, Task, Utils
|
2017-07-30 20:11:29 +02:00
|
|
|
from waflib.TaskGen import feature, after_method, taskgen_method
|
|
|
|
|
|
|
|
DONE = 0
|
|
|
|
DIRTY = 1
|
|
|
|
NEEDED = 2
|
|
|
|
|
2017-08-02 00:22:33 +02:00
|
|
|
SKIPPABLE = ['cshlib', 'cxxshlib', 'cstlib', 'cxxstlib', 'cprogram', 'cxxprogram']
|
|
|
|
|
2017-07-30 20:11:29 +02:00
|
|
|
class bld(Build.BuildContext):
|
|
|
|
def store(self):
|
2017-08-02 01:23:32 +02:00
|
|
|
# For each task generator, record all files involved in task objects
|
|
|
|
# optimization: done only if there was something built
|
|
|
|
for g in self.groups:
|
|
|
|
for tg in g:
|
|
|
|
do_cache = False
|
|
|
|
for tsk in tg.tasks:
|
|
|
|
if tsk.hasrun == Task.SUCCESS:
|
|
|
|
do_cache = True
|
|
|
|
pass
|
|
|
|
elif tsk.hasrun == Task.SKIPPED:
|
|
|
|
pass
|
2017-07-30 20:11:29 +02:00
|
|
|
else:
|
2017-08-02 01:23:32 +02:00
|
|
|
# dependencies are incomplete, clear the cache
|
|
|
|
try:
|
|
|
|
del self.raw_deps[(tg.path.abspath(), tg.idx)]
|
|
|
|
except KeyError:
|
|
|
|
pass
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
if not do_cache:
|
|
|
|
try:
|
|
|
|
self.raw_deps[(tg.path.abspath(), tg.idx)]
|
|
|
|
except KeyError:
|
|
|
|
# probably cleared because a wscript file changed
|
|
|
|
do_cache = True
|
2017-08-02 00:22:33 +02:00
|
|
|
|
2017-08-02 01:23:32 +02:00
|
|
|
if do_cache:
|
|
|
|
st = set()
|
|
|
|
for tsk in tg.tasks:
|
|
|
|
st.update(tsk.inputs)
|
|
|
|
st.update(self.node_deps.get(tsk.uid(), []))
|
|
|
|
|
|
|
|
lst = [x.abspath() for x in tg.path.ant_glob('wscript*')]
|
|
|
|
lst.extend(sorted(x.abspath() for x in st))
|
|
|
|
tss = [os.stat(x).st_mtime for x in lst]
|
|
|
|
self.raw_deps[(tg.path.abspath(), tg.idx)] = (lst, tss)
|
|
|
|
|
|
|
|
return Build.BuildContext.store(self)
|
2017-07-30 20:11:29 +02:00
|
|
|
|
|
|
|
def compute_needed_tgs(self):
|
|
|
|
# assume the 'use' keys are not modified during the build phase
|
|
|
|
|
|
|
|
# 1. obtain task generators that contain rebuilds
|
2017-08-02 00:22:33 +02:00
|
|
|
# 2. obtain the 'use' graph and its dual
|
2017-07-30 20:11:29 +02:00
|
|
|
stales = set()
|
|
|
|
reverse_use_map = Utils.defaultdict(list)
|
|
|
|
use_map = Utils.defaultdict(list)
|
|
|
|
|
|
|
|
for g in self.groups:
|
|
|
|
for tg in g:
|
2017-08-02 00:22:33 +02:00
|
|
|
if tg.is_stale():
|
|
|
|
stales.add(tg)
|
|
|
|
|
2017-07-30 20:11:29 +02:00
|
|
|
try:
|
|
|
|
lst = tg.use = Utils.to_list(tg.use)
|
|
|
|
except AttributeError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
for x in lst:
|
|
|
|
try:
|
|
|
|
xtg = self.get_tgen_by_name(x)
|
|
|
|
except Errors.WafError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
use_map[tg].append(xtg)
|
|
|
|
reverse_use_map[xtg].append(tg)
|
|
|
|
|
2017-08-02 00:22:33 +02:00
|
|
|
Logs.debug('rev_use: found %r stale tgs', len(stales))
|
|
|
|
|
2017-07-30 20:11:29 +02:00
|
|
|
# 3. dfs to post downstream tg as stale
|
|
|
|
visited = set()
|
|
|
|
def mark_down(tg):
|
|
|
|
if tg in visited:
|
|
|
|
return
|
|
|
|
visited.add(tg)
|
|
|
|
Logs.debug('rev_use: marking down %r as stale', tg.name)
|
|
|
|
tg.staleness = DIRTY
|
|
|
|
for x in reverse_use_map[tg]:
|
|
|
|
mark_down(x)
|
|
|
|
for tg in stales:
|
|
|
|
mark_down(tg)
|
|
|
|
|
|
|
|
# 4. dfs to find ancestors tg to mark as needed
|
|
|
|
self.needed_tgs = needed_tgs = set()
|
|
|
|
def mark_needed(tg):
|
|
|
|
if tg in needed_tgs:
|
|
|
|
return
|
|
|
|
needed_tgs.add(tg)
|
|
|
|
if tg.staleness == DONE:
|
|
|
|
Logs.debug('rev_use: marking up %r as needed', tg.name)
|
|
|
|
tg.staleness = NEEDED
|
|
|
|
for x in use_map[tg]:
|
|
|
|
mark_needed(x)
|
|
|
|
for xx in visited:
|
|
|
|
mark_needed(xx)
|
|
|
|
|
|
|
|
# so we have the whole tg trees to post in the set "needed"
|
|
|
|
# the stale ones should be fully build, while the needed ones
|
|
|
|
# may skip a few tasks, see create_compiled_task and apply_link_after below
|
|
|
|
Logs.debug('rev_use: amount of needed task gens: %r', len(needed_tgs))
|
|
|
|
|
|
|
|
def post_group(self):
|
|
|
|
# assumption: we can ignore the folder/subfolders cuts
|
|
|
|
def tgpost(tg):
|
|
|
|
try:
|
|
|
|
f = tg.post
|
|
|
|
except AttributeError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
f()
|
|
|
|
|
|
|
|
if not self.targets or self.targets == '*':
|
|
|
|
for tg in self.groups[self.current_group]:
|
|
|
|
# this can cut quite a lot of tg objects
|
|
|
|
if tg in self.needed_tgs:
|
|
|
|
tgpost(tg)
|
|
|
|
else:
|
|
|
|
# default implementation
|
|
|
|
return Build.BuildContext.post_group()
|
|
|
|
|
|
|
|
def get_build_iterator(self):
|
|
|
|
if not self.targets or self.targets == '*':
|
|
|
|
self.compute_needed_tgs()
|
|
|
|
return Build.BuildContext.get_build_iterator(self)
|
|
|
|
|
|
|
|
@taskgen_method
|
|
|
|
def is_stale(self):
|
|
|
|
# assume no globs
|
|
|
|
self.staleness = DIRTY
|
2017-08-02 00:22:33 +02:00
|
|
|
|
2017-08-02 01:23:32 +02:00
|
|
|
# 1. the case of always stale targets
|
2017-08-02 00:22:33 +02:00
|
|
|
if getattr(self, 'always_stale', False):
|
|
|
|
return True
|
|
|
|
|
2017-08-02 01:23:32 +02:00
|
|
|
# 2. check if the db file exists
|
|
|
|
db = os.path.join(self.bld.variant_dir, Context.DBFILE)
|
2017-07-30 20:11:29 +02:00
|
|
|
try:
|
|
|
|
dbstat = os.stat(db).st_mtime
|
|
|
|
except OSError:
|
2017-08-02 01:23:32 +02:00
|
|
|
Logs.debug('rev_use: must post %r because this is a clean build')
|
2017-07-30 20:11:29 +02:00
|
|
|
return True
|
|
|
|
|
2017-08-02 01:23:32 +02:00
|
|
|
# 3. check if the configuration changed
|
|
|
|
if os.stat(self.bld.bldnode.find_node('c4che/build.config.py').abspath()).st_mtime > dbstat:
|
2017-07-30 20:11:29 +02:00
|
|
|
Logs.debug('rev_use: must post %r because the configuration has changed', self.name)
|
|
|
|
return True
|
|
|
|
|
2017-08-02 01:23:32 +02:00
|
|
|
# 4. check if this is the first build (no cache)
|
2017-07-30 20:11:29 +02:00
|
|
|
try:
|
2017-08-02 01:23:32 +02:00
|
|
|
lst, tss = self.bld.raw_deps[(self.path.abspath(), self.idx)]
|
2017-07-30 20:11:29 +02:00
|
|
|
except KeyError:
|
2017-08-02 01:23:32 +02:00
|
|
|
Logs.debug('rev_use: must post %r because there it has no cached data', self.name)
|
2017-07-30 20:11:29 +02:00
|
|
|
return True
|
|
|
|
|
|
|
|
|
2017-08-02 00:22:33 +02:00
|
|
|
try:
|
|
|
|
cache = self.bld.cache_tstamp_rev_use
|
|
|
|
except AttributeError:
|
|
|
|
cache = self.bld.cache_tstamp_rev_use = {}
|
|
|
|
|
2017-07-30 20:11:29 +02:00
|
|
|
def tstamp(x):
|
2017-08-02 00:22:33 +02:00
|
|
|
# compute files timestamps with some caching
|
2017-07-30 20:11:29 +02:00
|
|
|
try:
|
|
|
|
return cache[x]
|
|
|
|
except KeyError:
|
|
|
|
ret = cache[x] = os.stat(x).st_mtime
|
|
|
|
return ret
|
|
|
|
|
2017-08-02 01:23:32 +02:00
|
|
|
# 5. check the timestamp of each dependency files listed is unchanged
|
|
|
|
for x, old_ts in zip(lst, tss):
|
2017-07-30 20:11:29 +02:00
|
|
|
try:
|
|
|
|
ts = tstamp(x)
|
|
|
|
except OSError:
|
2017-08-02 20:05:00 +02:00
|
|
|
del self.bld.raw_deps[(self.path.abspath(), self.idx)]
|
2017-07-30 20:11:29 +02:00
|
|
|
Logs.debug('rev_use: must post %r because %r does not exist anymore', self.name, x)
|
|
|
|
return True
|
|
|
|
else:
|
2017-08-02 01:23:32 +02:00
|
|
|
if ts != old_ts:
|
|
|
|
Logs.debug('rev_use: must post %r because the timestamp on %r changed %r %r', self.name, x, old_ts, ts)
|
2017-07-30 20:11:29 +02:00
|
|
|
return True
|
|
|
|
|
|
|
|
self.staleness = DONE
|
|
|
|
return False
|
|
|
|
|
|
|
|
@taskgen_method
|
|
|
|
def create_compiled_task(self, name, node):
|
2017-08-02 00:22:33 +02:00
|
|
|
# the purpose is to skip the creation of object files
|
2017-07-30 20:11:29 +02:00
|
|
|
if self.staleness == NEEDED:
|
2017-08-02 00:22:33 +02:00
|
|
|
# only libraries/programs can skip object files
|
|
|
|
for x in SKIPPABLE:
|
|
|
|
if x in self.features:
|
|
|
|
return None
|
2017-07-30 20:11:29 +02:00
|
|
|
|
|
|
|
out = '%s.%d.o' % (node.name, self.idx)
|
|
|
|
task = self.create_task(name, node, node.parent.find_or_declare(out))
|
|
|
|
try:
|
|
|
|
self.compiled_tasks.append(task)
|
|
|
|
except AttributeError:
|
|
|
|
self.compiled_tasks = [task]
|
|
|
|
return task
|
|
|
|
|
2017-08-02 00:22:33 +02:00
|
|
|
@feature(*SKIPPABLE)
|
2017-07-30 20:11:29 +02:00
|
|
|
@after_method('apply_link')
|
|
|
|
def apply_link_after(self):
|
2017-08-02 00:22:33 +02:00
|
|
|
# assumption: object-only targets are not skippable
|
|
|
|
# cprogram/cxxprogram might be unnecessary
|
2017-07-30 20:11:29 +02:00
|
|
|
if self.staleness != NEEDED:
|
|
|
|
return
|
|
|
|
try:
|
|
|
|
link_task = self.link_task
|
|
|
|
except AttributeError:
|
|
|
|
pass
|
|
|
|
else:
|
|
|
|
link_task.hasrun = Task.SKIPPED
|
|
|
|
|