diff --git a/ghost/Dockerfile b/ghost/Dockerfile deleted file mode 100644 index 35c5ca2..0000000 --- a/ghost/Dockerfile +++ /dev/null @@ -1,19 +0,0 @@ -FROM ubuntu:14.04 -MAINTAINER kev - -WORKDIR /tmp -ADD https://deb.nodesource.com/setup /tmp/ -ADD https://ghost.org/zip/ghost-latest.zip /tmp/ -RUN bash setup -RUN apt-get install -y nodejs unzip -RUN unzip ghost-latest.zip -d /ghost - -WORKDIR /ghost -RUN sed -e 's@127.0.0.1@0.0.0.0@' config.example.js > config.js -RUN npm install --production - -ENV NODE_ENV production -VOLUME /ghost -EXPOSE 2368 - -CMD npm start diff --git a/ghost/docker-compose.yml b/ghost/docker-compose.yml deleted file mode 100644 index 36aeb9c..0000000 --- a/ghost/docker-compose.yml +++ /dev/null @@ -1,5 +0,0 @@ -ghost: - image: ghost:latest - ports: - - "9000:2368" - restart: always diff --git a/scrapyd.old/Dockerfile b/scrapyd.old/Dockerfile deleted file mode 100644 index 860b687..0000000 --- a/scrapyd.old/Dockerfile +++ /dev/null @@ -1,58 +0,0 @@ -# -# Dockerfile for scrapyd -# - -FROM debian:jessie -MAINTAINER kev - -ADD ./scrapyd.conf /etc/scrapyd/ -ADD ./requirements.txt /etc/scrapyd/ - -RUN apt-get update \ - && apt-get install -y autoconf \ - build-essential \ - curl \ - git \ - libffi-dev \ - libssl-dev \ - libtool \ - python \ - python-dev \ - vim-tiny \ - && mkdir libxml2 \ - && curl -sSL ftp://xmlsoft.org/libxml2/libxml2-2.9.2.tar.gz | tar xz --strip 1 -C libxml2 \ - && cd libxml2 \ - && ./configure --prefix=/usr \ - && make \ - && make install \ - && ldconfig \ - && cd .. \ - && rm -rf libxml2 \ - && mkdir libxslt \ - && curl -sSL https://git.gnome.org/browse/libxslt/snapshot/libxslt-1.1.28.tar.xz | tar xJ --strip 1 -C libxslt \ - && cd libxslt \ - && ./autogen.sh --prefix=/usr \ - && make \ - && make install \ - && ldconfig \ - && cd .. \ - && rm -rf libxslt \ - && curl -sSL https://bootstrap.pypa.io/get-pip.py | python \ - && pip install scrapy==0.24.6 scrapyd==1.0.1 \ - && pip install -r /etc/scrapyd/requirements.txt \ - && curl -sSL https://github.com/scrapy/scrapy/raw/master/extras/scrapy_bash_completion -o /etc/bash_completion.d/scrapy_bash_completion \ - && echo 'source /etc/bash_completion.d/scrapy_bash_completion' >> /root/.bashrc \ - && apt-get remove -y autoconf \ - build-essential \ - libffi-dev \ - libssl-dev \ - libtool \ - python-dev \ - && apt-get autoremove -y \ - && rm -rf /var/lib/apt/lists/* - -VOLUME /etc/scrapyd/ /var/lib/scrapyd/ - -EXPOSE 6800 - -CMD ["scrapyd"] diff --git a/scrapyd.old/Dockerfile.old b/scrapyd.old/Dockerfile.old deleted file mode 100644 index ae3a837..0000000 --- a/scrapyd.old/Dockerfile.old +++ /dev/null @@ -1,42 +0,0 @@ -# -# Dockerfile for scrapyd -# -# References: -# - http://docs.docker.com/reference/builder/ -# - http://doc.scrapy.org/en/latest/topics/ubuntu.html#topics-ubuntu -# - https://github.com/scrapy/scrapyd/blob/master/debian/scrapyd.upstart#L9-L11 -# - http://pip.readthedocs.org/en/latest/installing.html -# - http://supervisord.org/index.html -# - -FROM ubuntu:14.04 -MAINTAINER kev - -ADD ./requirements.txt /tmp/ - -RUN apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv 627220E7 \ - && echo 'deb http://archive.scrapy.org/ubuntu scrapy main' > /etc/apt/sources.list.d/scrapy.list \ - && apt-get update \ - && apt-get install -y git \ - libpq-dev \ - build-essential \ - python-dev \ - python-pip \ - python-numpy \ - python-txzmq \ - scrapy-0.24 \ - scrapyd \ - supervisor \ - && pip install -r /tmp/requirements.txt \ - && apt-get remove -y build-essential \ - && apt-get autoremove -y \ - && apt-get clean -y \ - && rm -rf /var/lib/apt/lists/* - -ADD ./001-scrapyd /etc/scrapyd/conf.d/ -ADD ./scrapyd.conf /etc/supervisor/conf.d/ - -EXPOSE 6800 9001 - -CMD supervisord -n -c /etc/supervisor/supervisord.conf - diff --git a/scrapyd.old/README.md b/scrapyd.old/README.md deleted file mode 100644 index 212fa23..0000000 --- a/scrapyd.old/README.md +++ /dev/null @@ -1,15 +0,0 @@ -docker-scrapyd -============== - -Dockerfile for building an image that runs [scrapyd][1]. - -## Building - - $ docker build -t scrapyd . - -## Running - - $ docker run -p 6800:6800 scrapyd - -[1]: https://github.com/scrapy/scrapyd - diff --git a/scrapyd.old/requirements.txt b/scrapyd.old/requirements.txt deleted file mode 100644 index 6418dc4..0000000 --- a/scrapyd.old/requirements.txt +++ /dev/null @@ -1,21 +0,0 @@ ---allow-all-external ---allow-unverified jsonpath - -# parser -git+https://github.com/scrapy/scrapely -jsonpath -jsonschema -pyparsing -pyquery -pyyaml - -# database -pymongo -redis - -# others -chardet -requests -service_identity -toolz - diff --git a/scrapyd.old/scrapyd.conf b/scrapyd.old/scrapyd.conf deleted file mode 100644 index 77f31f5..0000000 --- a/scrapyd.old/scrapyd.conf +++ /dev/null @@ -1,26 +0,0 @@ -[scrapyd] -eggs_dir = /var/lib/scrapyd/eggs -logs_dir = /var/lib/scrapyd/logs -items_dir = /var/lib/scrapyd/items -dbs_dir = /var/lib/scrapyd/dbs -jobs_to_keep = 5 -max_proc = 0 -max_proc_per_cpu = 4 -finished_to_keep = 100 -poll_interval = 5 -http_port = 6800 -debug = off -runner = scrapyd.runner -application = scrapyd.app.application -launcher = scrapyd.launcher.Launcher - -[services] -schedule.json = scrapyd.webservice.Schedule -cancel.json = scrapyd.webservice.Cancel -addversion.json = scrapyd.webservice.AddVersion -listprojects.json = scrapyd.webservice.ListProjects -listversions.json = scrapyd.webservice.ListVersions -listspiders.json = scrapyd.webservice.ListSpiders -delproject.json = scrapyd.webservice.DeleteProject -delversion.json = scrapyd.webservice.DeleteVersion -listjobs.json = scrapyd.webservice.ListJobs