2563
Comment:
|
4734
|
Deletions are marked like this. | Additions are marked like this. |
Line 44: | Line 44: |
* clone my [[http://code.google.com/r/jasongrout-db-compute/|simple-db-compute]] repository (either just clone it locally, or clone it on google code and then pull from your clone) | * clone the [[http://code.google.com/p/simple-python-db-compute/|simple-db-compute]] repository (either just clone it locally, or clone it on google code and then pull from your clone) |
Line 49: | Line 49: |
* Look at making the device more parallel/scalable. See [[http://docs.python.org/library/multiprocessing.html|multiprocessing]], or maybe use the [[http://hg.sagemath.org/sage-main/file/tip/sage/parallel|parallel code from Sage]]. | * Look at making the device more parallel/scalable. See [[http://docs.python.org/library/multiprocessing.html|multiprocessing]] (which includes functionality for pools of worker processes), or maybe use the [[http://hg.sagemath.org/sage-main/file/tip/sage/parallel|parallel code from Sage]]. The new experiments in [[https://groups.google.com/forum/#!topic/sage-devel/64dI4v_wtz0|forking Sage to start it up]] also seem relevant. |
Line 51: | Line 51: |
== Projects == Here are some project ideas for simple-db-compute, along with some hopefully helpful pointers to resources. * Make the web interface use AJAX to send a computation and display a result. Helpful resources: [[http://api.jquery.com/jQuery.ajax/|Jquery AJAX]], google for numerous jquery ajax tutorials. You will probably want to create a javascript file in the static/ directory and add it to templates/base.html (follow the examples already there adding jquery, for example). I would suggest using something like JSON to send and receive messages with the server. * Make flask assign a random computation id to each incoming request, return that computation id to the browser as a return value in the ajax call. Then make the browser keep asking the server for the results of that particular computation id. When results exist, send the result back to the browser. Otherwise, send back some message that the results are not computed yet. * Write a test script that will hammer the site hard to test how scalable it is. Maybe the python multiprocessing module could be used to make a number of workers in a pool and each worker submit computations to the website at a configurable rate. Record the time for roundtrip in a request and the time for a computation to appear. See if the site scales to hundreds of requests each minute or so. * On the backend, write a device that keeps a pool of workers (maybe using the python multiprocessing library) and keeps those workers busy with computations from the database. Ideally, the polling of the database should not be blocked by worker computations. Instead, on each poll, workers that are finished should have output put into the database, and new computations should be pulled out of the database. It seems like a good idea to avoid trying to put each output in the database as it happens. Rather, batch up the database updates to happen once in a polling interval. |
Drake Sage Group
This page documents activities of the Drake University Sage group.
Our initial work is on a single-cell compute server, which basically is a webpage that can execute an arbitrary block of Sage code.
For more information, please contact Jason Grout at jason#[email protected] (replace the # with a .)
03 Feb 2010
Meet in Howard Hall 308 at 2pm (room reserved from 1:30-3, so come early if you want).
Agenda
- Introductions
- What Sage is
- Overview of the simple-db-compute project and its architecture
- Resources
Sage Notebook mailing list -- subscribe to this
- Installfest--get the simple compute server up and running on as many people's computers as possible
install mongodb
install PyMongo and Flask python modules:
# from within python from setuptools.command import easy_install easy_install.main(["flask"]) easy_install.main(["pymongo"])
configure mercurial: put this in your ~/.hgrc file
[ui] username = YOUR NAME <YOUR EMAIL> [extensions] record= convert= hgext.mq= hgext.extdiff= hgk= transplant= fetch=
Create a Google code account
clone the simple-db-compute repository (either just clone it locally, or clone it on google code and then pull from your clone)
- First goal of project
- familiarize yourself with the simple-db-compute source code
- add a "compute id" that is returned to the user. The answers page then queries for just that computation's result.
- add necessary files to get this running on Windows (for example, a .bat file to start mongodb)
Look at making the device more parallel/scalable. See multiprocessing (which includes functionality for pools of worker processes), or maybe use the parallel code from Sage. The new experiments in forking Sage to start it up also seem relevant.
- Next meeting time
Projects
Here are some project ideas for simple-db-compute, along with some hopefully helpful pointers to resources.
Make the web interface use AJAX to send a computation and display a result. Helpful resources: Jquery AJAX, google for numerous jquery ajax tutorials. You will probably want to create a javascript file in the static/ directory and add it to templates/base.html (follow the examples already there adding jquery, for example). I would suggest using something like JSON to send and receive messages with the server.
- Make flask assign a random computation id to each incoming request, return that computation id to the browser as a return value in the ajax call. Then make the browser keep asking the server for the results of that particular computation id. When results exist, send the result back to the browser. Otherwise, send back some message that the results are not computed yet.
- Write a test script that will hammer the site hard to test how scalable it is. Maybe the python multiprocessing module could be used to make a number of workers in a pool and each worker submit computations to the website at a configurable rate. Record the time for roundtrip in a request and the time for a computation to appear. See if the site scales to hundreds of requests each minute or so.
- On the backend, write a device that keeps a pool of workers (maybe using the python multiprocessing library) and keeps those workers busy with computations from the database. Ideally, the polling of the database should not be blocked by worker computations. Instead, on each poll, workers that are finished should have output put into the database, and new computations should be pulled out of the database. It seems like a good idea to avoid trying to put each output in the database as it happens. Rather, batch up the database updates to happen once in a polling interval.