2
0
mirror of https://github.com/vimagick/dockerfiles.git synced 2025-12-21 21:33:02 +01:00

delete ghost

This commit is contained in:
kev
2015-06-17 14:14:45 +08:00
parent 5862b9f29d
commit 50c030a5af
7 changed files with 0 additions and 186 deletions

View File

@@ -1,19 +0,0 @@
FROM ubuntu:14.04
MAINTAINER kev
WORKDIR /tmp
ADD https://deb.nodesource.com/setup /tmp/
ADD https://ghost.org/zip/ghost-latest.zip /tmp/
RUN bash setup
RUN apt-get install -y nodejs unzip
RUN unzip ghost-latest.zip -d /ghost
WORKDIR /ghost
RUN sed -e 's@127.0.0.1@0.0.0.0@' config.example.js > config.js
RUN npm install --production
ENV NODE_ENV production
VOLUME /ghost
EXPOSE 2368
CMD npm start

View File

@@ -1,5 +0,0 @@
ghost:
image: ghost:latest
ports:
- "9000:2368"
restart: always

View File

@@ -1,58 +0,0 @@
#
# Dockerfile for scrapyd
#
FROM debian:jessie
MAINTAINER kev <noreply@datageek.info>
ADD ./scrapyd.conf /etc/scrapyd/
ADD ./requirements.txt /etc/scrapyd/
RUN apt-get update \
&& apt-get install -y autoconf \
build-essential \
curl \
git \
libffi-dev \
libssl-dev \
libtool \
python \
python-dev \
vim-tiny \
&& mkdir libxml2 \
&& curl -sSL ftp://xmlsoft.org/libxml2/libxml2-2.9.2.tar.gz | tar xz --strip 1 -C libxml2 \
&& cd libxml2 \
&& ./configure --prefix=/usr \
&& make \
&& make install \
&& ldconfig \
&& cd .. \
&& rm -rf libxml2 \
&& mkdir libxslt \
&& curl -sSL https://git.gnome.org/browse/libxslt/snapshot/libxslt-1.1.28.tar.xz | tar xJ --strip 1 -C libxslt \
&& cd libxslt \
&& ./autogen.sh --prefix=/usr \
&& make \
&& make install \
&& ldconfig \
&& cd .. \
&& rm -rf libxslt \
&& curl -sSL https://bootstrap.pypa.io/get-pip.py | python \
&& pip install scrapy==0.24.6 scrapyd==1.0.1 \
&& pip install -r /etc/scrapyd/requirements.txt \
&& curl -sSL https://github.com/scrapy/scrapy/raw/master/extras/scrapy_bash_completion -o /etc/bash_completion.d/scrapy_bash_completion \
&& echo 'source /etc/bash_completion.d/scrapy_bash_completion' >> /root/.bashrc \
&& apt-get remove -y autoconf \
build-essential \
libffi-dev \
libssl-dev \
libtool \
python-dev \
&& apt-get autoremove -y \
&& rm -rf /var/lib/apt/lists/*
VOLUME /etc/scrapyd/ /var/lib/scrapyd/
EXPOSE 6800
CMD ["scrapyd"]

View File

@@ -1,42 +0,0 @@
#
# Dockerfile for scrapyd
#
# References:
# - http://docs.docker.com/reference/builder/
# - http://doc.scrapy.org/en/latest/topics/ubuntu.html#topics-ubuntu
# - https://github.com/scrapy/scrapyd/blob/master/debian/scrapyd.upstart#L9-L11
# - http://pip.readthedocs.org/en/latest/installing.html
# - http://supervisord.org/index.html
#
FROM ubuntu:14.04
MAINTAINER kev <noreply@datageek.info>
ADD ./requirements.txt /tmp/
RUN apt-key adv --keyserver hkp://keyserver.ubuntu.com:80 --recv 627220E7 \
&& echo 'deb http://archive.scrapy.org/ubuntu scrapy main' > /etc/apt/sources.list.d/scrapy.list \
&& apt-get update \
&& apt-get install -y git \
libpq-dev \
build-essential \
python-dev \
python-pip \
python-numpy \
python-txzmq \
scrapy-0.24 \
scrapyd \
supervisor \
&& pip install -r /tmp/requirements.txt \
&& apt-get remove -y build-essential \
&& apt-get autoremove -y \
&& apt-get clean -y \
&& rm -rf /var/lib/apt/lists/*
ADD ./001-scrapyd /etc/scrapyd/conf.d/
ADD ./scrapyd.conf /etc/supervisor/conf.d/
EXPOSE 6800 9001
CMD supervisord -n -c /etc/supervisor/supervisord.conf

View File

@@ -1,15 +0,0 @@
docker-scrapyd
==============
Dockerfile for building an image that runs [scrapyd][1].
## Building
$ docker build -t scrapyd .
## Running
$ docker run -p 6800:6800 scrapyd
[1]: https://github.com/scrapy/scrapyd

View File

@@ -1,21 +0,0 @@
--allow-all-external
--allow-unverified jsonpath
# parser
git+https://github.com/scrapy/scrapely
jsonpath
jsonschema
pyparsing
pyquery
pyyaml
# database
pymongo
redis
# others
chardet
requests
service_identity
toolz

View File

@@ -1,26 +0,0 @@
[scrapyd]
eggs_dir = /var/lib/scrapyd/eggs
logs_dir = /var/lib/scrapyd/logs
items_dir = /var/lib/scrapyd/items
dbs_dir = /var/lib/scrapyd/dbs
jobs_to_keep = 5
max_proc = 0
max_proc_per_cpu = 4
finished_to_keep = 100
poll_interval = 5
http_port = 6800
debug = off
runner = scrapyd.runner
application = scrapyd.app.application
launcher = scrapyd.launcher.Launcher
[services]
schedule.json = scrapyd.webservice.Schedule
cancel.json = scrapyd.webservice.Cancel
addversion.json = scrapyd.webservice.AddVersion
listprojects.json = scrapyd.webservice.ListProjects
listversions.json = scrapyd.webservice.ListVersions
listspiders.json = scrapyd.webservice.ListSpiders
delproject.json = scrapyd.webservice.DeleteProject
delversion.json = scrapyd.webservice.DeleteVersion
listjobs.json = scrapyd.webservice.ListJobs