| | from dreamcoder.likelihoodModel import AllOrNothingLikelihoodModel |
| | from dreamcoder.grammar import * |
| | from dreamcoder.utilities import get_root_dir |
| |
|
| | import os |
| | import traceback |
| | import subprocess |
| |
|
| |
|
| | def multicoreEnumeration(g, tasks, _=None, |
| | enumerationTimeout=None, |
| | solver='ocaml', |
| | CPUs=1, |
| | maximumFrontier=None, |
| | verbose=True, |
| | evaluationTimeout=None, |
| | testing=False): |
| | '''g: Either a Grammar, or a map from task to grammar. |
| | Returns (list-of-frontiers, map-from-task-to-search-time)''' |
| |
|
| | |
| | |
| | |
| |
|
| | from multiprocessing import Queue |
| |
|
| | |
| | import dill |
| |
|
| | solvers = {"ocaml": solveForTask_ocaml, |
| | "pypy": solveForTask_pypy, |
| | "python": solveForTask_python} |
| | assert solver in solvers, "You must specify a valid solver. options are ocaml, pypy, or python." |
| |
|
| | likelihoodModel = None |
| | if solver == 'pypy' or solver == 'python': |
| | |
| | likelihoodModel = AllOrNothingLikelihoodModel(timeout=evaluationTimeout) |
| | |
| | solver = solvers[solver] |
| |
|
| | if not isinstance(g, dict): |
| | g = {t: g for t in tasks} |
| | task2grammar = g |
| |
|
| | |
| | |
| | |
| | |
| | |
| | jobs = {} |
| | for i, t in enumerate(tasks): |
| | if testing: |
| | k = (task2grammar[t], t.request, i) |
| | else: |
| | k = (task2grammar[t], t.request) |
| | jobs[k] = jobs.get(k, []) + [t] |
| |
|
| | disableParallelism = len(jobs) == 1 |
| | parallelCallback = launchParallelProcess if not disableParallelism else lambda f, * \ |
| | a, **k: f(*a, **k) |
| | if disableParallelism: |
| | eprint("Disabling parallelism on the Python side because we only have one job.") |
| | eprint("If you are using ocaml, there could still be parallelism.") |
| |
|
| | |
| | bestSearchTime = {t: None for t in task2grammar} |
| |
|
| | lowerBounds = {k: 0. for k in jobs} |
| |
|
| | frontiers = {t: Frontier([], task=t) for t in task2grammar} |
| |
|
| | |
| | stopwatches = {t: Stopwatch() for t in jobs} |
| |
|
| | |
| | taskToNumberOfPrograms = {t: 0 for t in tasks } |
| |
|
| | def numberOfHits(f): |
| | return sum(e.logLikelihood > -0.01 for e in f) |
| |
|
| | def budgetIncrement(lb): |
| | if True: |
| | return 1.5 |
| | |
| | if lb < 24.: |
| | return 1. |
| | elif lb < 27.: |
| | return 0.5 |
| | else: |
| | return 0.25 |
| |
|
| | def maximumFrontiers(j): |
| | tasks = jobs[j] |
| | return {t: maximumFrontier - numberOfHits(frontiers[t]) for t in tasks} |
| |
|
| | def allocateCPUs(n, tasks): |
| | allocation = {t: 0 for t in tasks} |
| | while n > 0: |
| | for t in tasks: |
| | |
| | if testing and allocation[t] > 0: |
| | return allocation |
| | allocation[t] += 1 |
| | n -= 1 |
| | if n == 0: |
| | break |
| | return allocation |
| |
|
| | def refreshJobs(): |
| | for k in list(jobs.keys()): |
| | v = [t for t in jobs[k] |
| | if numberOfHits(frontiers[t]) < maximumFrontier |
| | and stopwatches[k].elapsed <= enumerationTimeout] |
| | if v: |
| | jobs[k] = v |
| | else: |
| | del jobs[k] |
| |
|
| | |
| | q = Queue() |
| |
|
| | |
| | activeCPUs = 0 |
| |
|
| | |
| | id2CPUs = {} |
| | |
| | id2job = {} |
| | nextID = 0 |
| |
|
| | while True: |
| | refreshJobs() |
| | |
| | |
| | |
| | freeJobs = [j for j in jobs if not stopwatches[j].running |
| | and stopwatches[j].elapsed < enumerationTimeout - 0.5] |
| | if freeJobs and activeCPUs < CPUs: |
| | |
| | |
| | freeJobs.sort(key=lambda j: lowerBounds[j]) |
| | |
| | availableCPUs = CPUs - activeCPUs |
| | allocation = allocateCPUs(availableCPUs, freeJobs) |
| | for j in freeJobs: |
| | if allocation[j] == 0: |
| | continue |
| | g, request = j[:2] |
| | bi = budgetIncrement(lowerBounds[j]) |
| | thisTimeout = enumerationTimeout - stopwatches[j].elapsed |
| | eprint("(python) Launching %s (%d tasks) w/ %d CPUs. %f <= MDL < %f. Timeout %f." % |
| | (request, len(jobs[j]), allocation[j], lowerBounds[j], lowerBounds[j] + bi, thisTimeout)) |
| | stopwatches[j].start() |
| | parallelCallback(wrapInThread(solver), |
| | q=q, g=g, ID=nextID, |
| | elapsedTime=stopwatches[j].elapsed, |
| | CPUs=allocation[j], |
| | tasks=jobs[j], |
| | lowerBound=lowerBounds[j], |
| | upperBound=lowerBounds[j] + bi, |
| | budgetIncrement=bi, |
| | timeout=thisTimeout, |
| | evaluationTimeout=evaluationTimeout, |
| | maximumFrontiers=maximumFrontiers(j), |
| | testing=testing, |
| | likelihoodModel=likelihoodModel) |
| | id2CPUs[nextID] = allocation[j] |
| | id2job[nextID] = j |
| | nextID += 1 |
| |
|
| | activeCPUs += allocation[j] |
| | lowerBounds[j] += bi |
| |
|
| | |
| | |
| | if all(not s.running for s in stopwatches.values()): |
| | break |
| |
|
| | |
| | message = Bunch(dill.loads(q.get())) |
| |
|
| | if message.result == "failure": |
| | eprint("PANIC! Exception in child worker:", message.exception) |
| | eprint(message.stacktrace) |
| | assert False |
| | elif message.result == "success": |
| | |
| | activeCPUs -= id2CPUs[message.ID] |
| | stopwatches[id2job[message.ID]].stop() |
| |
|
| | newFrontiers, searchTimes, pc = message.value |
| | for t, f in newFrontiers.items(): |
| | oldBest = None if len( |
| | frontiers[t]) == 0 else frontiers[t].bestPosterior |
| | frontiers[t] = frontiers[t].combine(f) |
| | newBest = None if len( |
| | frontiers[t]) == 0 else frontiers[t].bestPosterior |
| |
|
| | taskToNumberOfPrograms[t] += pc |
| |
|
| | dt = searchTimes[t] |
| | if dt is not None: |
| | if bestSearchTime[t] is None: |
| | bestSearchTime[t] = dt |
| | else: |
| | |
| | assert oldBest is not None |
| | assert newBest is not None |
| | newScore = newBest.logPrior + newBest.logLikelihood |
| | oldScore = oldBest.logPrior + oldBest.logLikelihood |
| |
|
| | if newScore > oldScore: |
| | bestSearchTime[t] = dt |
| | elif newScore == oldScore: |
| | bestSearchTime[t] = min(bestSearchTime[t], dt) |
| | else: |
| | eprint("Unknown message result:", message.result) |
| | assert False |
| |
|
| | eprint("We enumerated this many programs, for each task:\n\t", |
| | list(taskToNumberOfPrograms.values())) |
| |
|
| | return [frontiers[t] for t in tasks], bestSearchTime |
| |
|
| | def wrapInThread(f): |
| | """ |
| | Returns a function that is designed to be run in a thread/threadlike process. |
| | Result will be either put into the q |
| | """ |
| | import dill |
| |
|
| | def _f(*a, **k): |
| | q = k.pop("q") |
| | ID = k.pop("ID") |
| |
|
| | try: |
| | r = f(*a, **k) |
| | q.put(dill.dumps({"result": "success", |
| | "ID": ID, |
| | "value": r})) |
| | except Exception as e: |
| | q.put(dill.dumps({"result": "failure", |
| | "exception": e, |
| | "stacktrace": traceback.format_exc(), |
| | "ID": ID})) |
| | return |
| | return _f |
| |
|
| |
|
| | def solveForTask_ocaml(_=None, |
| | elapsedTime=0., |
| | CPUs=1, |
| | g=None, tasks=None, |
| | lowerBound=None, upperBound=None, budgetIncrement=None, |
| | timeout=None, |
| | testing=None, |
| | likelihoodModel=None, |
| | evaluationTimeout=None, maximumFrontiers=None): |
| |
|
| | import json |
| |
|
| | def taskMessage(t): |
| | m = { |
| | "examples": [{"inputs": list(xs), "output": y} for xs, y in t.examples], |
| | "name": t.name, |
| | "request": t.request.json(), |
| | "maximumFrontier": maximumFrontiers[t]} |
| | if hasattr(t, "specialTask"): |
| | special, extra = t.specialTask |
| | m["specialTask"] = special |
| | m["extras"] = extra |
| | return m |
| |
|
| |
|
| | message = {"DSL": g.json(), |
| | "tasks": [taskMessage(t) |
| | for t in tasks], |
| |
|
| | "programTimeout": evaluationTimeout, |
| | "nc": CPUs, |
| | "timeout": timeout, |
| | "lowerBound": lowerBound, |
| | "upperBound": upperBound, |
| | "budgetIncrement": budgetIncrement, |
| | "verbose": False, |
| | "shatter": 5 if len(tasks) == 1 and "turtle" in str(tasks[0].request) else 10} |
| |
|
| | if hasattr(tasks[0], 'maxParameters') and tasks[0].maxParameters is not None: |
| | message["maxParameters"] = tasks[0].maxParameters |
| |
|
| | message = json.dumps(message) |
| | |
| | |
| |
|
| | try: |
| | solver_file = os.path.join(get_root_dir(), 'solver') |
| | process = subprocess.Popen(solver_file, |
| | stdin=subprocess.PIPE, |
| | stdout=subprocess.PIPE) |
| | response, error = process.communicate(bytes(message, encoding="utf-8")) |
| | response = json.loads(response.decode("utf-8")) |
| | except OSError as exc: |
| | raise exc |
| |
|
| | except: |
| | print("response:", response) |
| | print("error:", error) |
| | with open("message", "w") as f: |
| | f.write(message) |
| | print("message,", message) |
| | assert False, "MAX RAISE" |
| |
|
| |
|
| | pc = response.get("number_enumerated",0) |
| | frontiers = {} |
| | searchTimes = {} |
| | for t in tasks: |
| | solutions = response[t.name] |
| | frontier = Frontier([FrontierEntry(program=p, |
| | logLikelihood=e["logLikelihood"], |
| | logPrior=g.logLikelihood(t.request, p)) |
| | for e in solutions |
| | for p in [Program.parse(e["program"])]], |
| | task=t) |
| | frontiers[t] = frontier |
| | if frontier.empty: |
| | searchTimes[t] = None |
| | |
| | |
| | |
| | |
| | |
| | |
| | else: |
| | searchTimes[t] = min( |
| | (e["logLikelihood"] + e["logPrior"], |
| | e["time"]) for e in solutions)[1] + elapsedTime |
| |
|
| | return frontiers, searchTimes, pc |
| |
|
| | def solveForTask_pypy(_=None, |
| | elapsedTime=0., |
| | g=None, task=None, |
| | lowerBound=None, upperBound=None, budgetIncrement=None, |
| | timeout=None, |
| | likelihoodModel=None, |
| | evaluationTimeout=None, maximumFrontier=None, testing=False): |
| | return callCompiled(enumerateForTasks, |
| | g, tasks, likelihoodModel, |
| | timeout=timeout, |
| | testing=testing, |
| | elapsedTime=elapsedTime, |
| | evaluationTimeout=evaluationTimeout, |
| | maximumFrontiers=maximumFrontiers, |
| | budgetIncrement=budgetIncrement, |
| | lowerBound=lowerBound, upperBound=upperBound) |
| |
|
| | def solveForTask_python(_=None, |
| | elapsedTime=0., |
| | g=None, tasks=None, |
| | lowerBound=None, upperBound=None, budgetIncrement=None, |
| | timeout=None, |
| | CPUs=1, |
| | likelihoodModel=None, |
| | evaluationTimeout=None, maximumFrontiers=None, testing=False): |
| | return enumerateForTasks(g, tasks, likelihoodModel, |
| | timeout=timeout, |
| | testing=testing, |
| | elapsedTime=elapsedTime, |
| | evaluationTimeout=evaluationTimeout, |
| | maximumFrontiers=maximumFrontiers, |
| | budgetIncrement=budgetIncrement, |
| | lowerBound=lowerBound, upperBound=upperBound) |
| |
|
| |
|
| | class EnumerationTimeout(Exception): |
| | pass |
| |
|
| | def enumerateForTasks(g, tasks, likelihoodModel, _=None, |
| | verbose=False, |
| | timeout=None, |
| | elapsedTime=0., |
| | CPUs=1, |
| | testing=False, |
| | evaluationTimeout=None, |
| | lowerBound=0., |
| | upperBound=100., |
| | budgetIncrement=1.0, maximumFrontiers=None): |
| | assert timeout is not None, \ |
| | "enumerateForTasks: You must provide a timeout." |
| |
|
| | from time import time |
| |
|
| | request = tasks[0].request |
| | assert all(t.request == request for t in tasks), \ |
| | "enumerateForTasks: Expected tasks to all have the same type" |
| |
|
| | maximumFrontiers = [maximumFrontiers[t] for t in tasks] |
| | |
| | |
| | hits = [PQ() for _ in tasks] |
| |
|
| | starting = time() |
| | previousBudget = lowerBound |
| | budget = lowerBound + budgetIncrement |
| | try: |
| | totalNumberOfPrograms = 0 |
| | while time() < starting + timeout and \ |
| | any(len(h) < mf for h, mf in zip(hits, maximumFrontiers)) and \ |
| | budget <= upperBound: |
| | numberOfPrograms = 0 |
| |
|
| | for prior, _, p in g.enumeration(Context.EMPTY, [], request, |
| | maximumDepth=99, |
| | upperBound=budget, |
| | lowerBound=previousBudget): |
| | descriptionLength = -prior |
| | |
| | assert descriptionLength <= budget |
| | |
| | assert descriptionLength > previousBudget |
| |
|
| | numberOfPrograms += 1 |
| | totalNumberOfPrograms += 1 |
| |
|
| | for n in range(len(tasks)): |
| | task = tasks[n] |
| |
|
| | |
| | |
| | |
| | |
| | success, likelihood = likelihoodModel.score(p, task) |
| | if not success: |
| | continue |
| | |
| | dt = time() - starting + elapsedTime |
| | priority = -(likelihood + prior) |
| | hits[n].push(priority, |
| | (dt, FrontierEntry(program=p, |
| | logLikelihood=likelihood, |
| | logPrior=prior))) |
| | if len(hits[n]) > maximumFrontiers[n]: |
| | hits[n].popMaximum() |
| |
|
| | if timeout is not None and time() - starting > timeout: |
| | raise EnumerationTimeout |
| |
|
| | previousBudget = budget |
| | budget += budgetIncrement |
| |
|
| | if budget > upperBound: |
| | break |
| | except EnumerationTimeout: |
| | pass |
| | frontiers = {tasks[n]: Frontier([e for _, e in hits[n]], |
| | task=tasks[n]) |
| | for n in range(len(tasks))} |
| | searchTimes = { |
| | tasks[n]: None if len(hits[n]) == 0 else \ |
| | min(t for t,_ in hits[n]) for n in range(len(tasks))} |
| |
|
| | return frontiers, searchTimes, totalNumberOfPrograms |
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|