-
Notifications
You must be signed in to change notification settings - Fork 54
FAQ
- I get an error when I try to launch an experiment
- I'm running into Docker issues. How do I resolve these?
- Experiments hang at initialization. How do I fix this?
- How do I get many participants for some study?
- How do I specify the participant ID?
- How do I access all the targets?
- How do I restart a machine I stopped on EC2?
- How do I time every line in a function?
- How do I include feature vectors with targets?
- Launching an experiment takes a long time. How do I debug with this?
Specifically, this is seen:
$ python launch.py strange_fruit_triplet/init.yaml strange_fruit_triplet/strangefruit30.zip
# ...
An error occured launching the experiment:
The difference between the request time and the current time is too large.
Details:
Traceback (most recent call last):
# ...
Traceback (most recent call last):
# ...
ValueError: Experiment didn't launch successfully
To resolve this, restart docker. More is covered in this SO answer: https://stackoverflow.com/questions/27674968/amazon-s3-docker-403-forbidden-the-difference-between-the-request-time-and
Try removing all the docker machines and resetting to factory defaults. [A guide on this] (https://techoverflow.net/blog/2013/10/22/docker-remove-all-images-and-containers/) and the related GitHub issue says that
#!/bin/bash
# Delete all containers
docker rm $(docker ps -a -q)
# Delete all images
docker rmi $(docker images -q)
On Mac, I set Docker for Mac to factory defaults in Docker for Mac > Preferences > Reset. This also can be run using
Note: this will delete all Docker data! Be sure to backup and/or delete if you need to.
The issue: you launch an experiment (possibly through test_api.py
). But all that's printed to stdout
is Lauching experiment...
and the docker logs don't get past the initial printing. It appears to be hanging at the initialization of the experiment.
A method we've tried with some success has been
docker-compose stop # typically through cntrl-C
docker-compose rm
docker-compose up # or ./docker_up.sh
If you launched via the AMI, you can stop/start your machine. This will restart the docker machines fix this issue.
We have used Mechanical Turk. We setup some machine (and obtain the URL), then direct them to this URL. Here they answer 50 questions, with no interactions with MTurk. At the end, we ask them to copy-paste their User ID (shown by default at the end of the study) back into MTurk. Using this, we can verify that they responded.
Instead of going to [next-url]:8000/.../[exp-uid]
, go to [next-url]:8000/.../[exp-uid]?participant=[id]
.
e.g., instead of http://localhost:8000/query/query_page/query_page/368de69569286ce0ba8a3f40b58a2a
go to http://localhost:8000/query/query_page/query_page/368de69569286ce0ba8a3f40b58a2a?participant=scott
targets = butler.targets.get_targetset(butler.exp_uid)
- Did you launch via the
next_ec2.py
script?- Use the
next_ec2.py
script. It has commands forstart
andstop
, which mirror the EC2 GUI commands in the "Actions > Instate state" menu.
- Use the
- AMI launch
- The "Actions > Instate state" menu should work.
Please, make backups before stopping machines.
You add the lines
from decorator import decorator
from line_profiler import LineProfiler
@decorator
def profile_each_line(func, *args, **kwargs):
profiler = LineProfiler()
profiled_func = profiler(func)
retval = None
try:
retval = profiled_func(*args, **kwargs)
finally:
profiler.print_stats()
return retval
somewhere along with all your other imports at the beginning of the file, and then you can use @profile_each_line
as a decorator for any function you wish to profile and you will get many statistics about how long each line takes to run, and which lines took what percentage of the total runtime, etc.
There are three options:
- NEXT accepts a list of dictionaries as targets. These dictionaries get stored in the butler and are accessible. This requires launching the experiment yourself (and writing any necessary scripts).
- Enforcing that feature vectors be passed in to your app in
initExp
can be done in the YAML. This requires developing your own app. - We have also developed a feature to allow adding feature vectors to images to the examples in
example/
. The example below will illustrate adding feature vectors to an existing application, the primary empirical use case we have seen.
The third option in detail:
Advantages of this approach include using your algorithm with an existing application/framework. You can easily compare your algorithm with other algorithms. A new algorithm has a choice of paying attention to feature vectors or not; it's up to the developer of that algorithm.
We need to modify the file that launches the experiment on NEXT (e.g., examples/strangefruit/experiment_triplet.py
). In this, if we include a key target_features
in the experiment dictionary, feature vectors will be added by example/launch_experiment.py
(note: only for images ending in .png
or .jpg
).
The dictionary we add will have keys of different filenames and values of the feature vector. i.e.,
the dictionary is the form of {filename: feature_vector}
. We
experiment['primary_type'] = 'image'
target_zip = 'strangefruit30.zip'
experiment['primary_target_file'] = target_zip
experiment['target_features'] = {filename.split('/')[-1]: np.random.rand(2).tolist() # tolist() because numpy array not serializable
for filename in zipfile.ZipFile(target_zip).namelist()}
# filename.split above removes 'strangefuit/' from 'strangefruit/image.png'. Required for
# use of lauch_experiment.py (which the examples in next/examples use)
Note: This is only provides information on putting features in targets. It not give information on how to load feature vectors (although I would use np.load
or scipy.io.loadmat
).
Then to access these in myAlg.py
, in initExp
we include these lines:
import numpy as np
class myAlg:
def initExp(self, butler, ...):
targets = butler.targets.get_targetset(butler.exp_uid)
feature_matrix = [target['feature_vector'] for target in targets]
feature_matrix = np.array(feature_matrix)
# ...
Do not run docker-compose rm
, as it removes your containers. If you run docker-compose stop; docker-compose start
, your experiment will remain (docker-compose stop
is typically run via Cntrl-C). For more detail, see the wiki page on debugging.