一些理論知識,我這裏就不累贅了html
docker 入門資料,參考:https://yeasy.gitbooks.io/docker_practice/content/前端
Dockerfile經常使用命令,圖片來源於網絡node
Dockerfile 打包控制檯應用程序mysql
新建一個控制檯程序,控制檯程序添加一個文本文件,去掉.txt 擴展名,改爲Dockerfile 輸入如下代碼linux
FROM microsoft/dotnet:sdk AS build WORKDIR /code COPY *.csproj /code RUN dotnet restore COPY . /code RUN dotnet publish -c Release -o out FROM microsoft/dotnet:runtime WORKDIR /app COPY --from=build /code/out /app ENTRYPOINT ["dotnet","console.dll"]
Program.cs 中編寫測試代碼git
一切準備完成。就是build把項目打包成鏡像了github
切換到當前項目路徑下。輸入: docker build -t cn/console:v1 . web
docker build -t :是打包固有的命令redis
cn/console:v1 :sql
cn:是組織名稱或者說是用戶名,若是你想把本身的鏡像push到hub.docker 上,cn必須是你本身的用戶名
console:是鏡像名稱
v1:是tag。一個標籤,能夠用來區分同一個鏡像,不一樣用途。,若是不指定。默認是latest
. :表明當前目錄爲上下文,dockerfile也一定是在當前目錄下
回車後,會看到一系列的執行步驟,dockerfile中。一條命令就是一個步驟
經過 docker images 能夠查看全部鏡像
經過docker images cn/console 查看相關鏡像
好比我本地有3個 cn/console鏡像,但tag不一樣
既然鏡像有了。那麼就能夠根據鏡像生成容器了。容器是鏡像的一個實例。鏡像運行起來纔會有容器,就跟類和對象同樣,new一個類,是實例化的操做
輸入命令:
docker run --name myfirst cn/console:v1
由於是佔用前端線程運行容器,全部界面沒法繼續輸入命令了。能夠Ctrl+c 結束容器運行
從上面的dockerfile。你會發現,咱們是把源碼打包成鏡像的。也就是執行了restore,到Release操做
其實若是你是已經Release後的文件了。dockerfile能夠更簡單
FROM microsoft/dotnet WORKDIR /app COPY . /app CMD ["dotnet","run"]
以上就是一個基礎的程序打包成鏡像,我以爲這不是重點,經常使用的應該是應用程序,而不是控制檯程序
後面打算把net core api打包成鏡像。在講這個以前,咱們先來搭建好環境。
由於我有個阿里雲服務器(CentOS7),而後有2檯筆記本,一個是Docker for Windows 環境,一個是CentOS7,因此常常會在這3個環境中來回折騰
兩種系統仍是有區別的,至少我弄的時候,遇到過很多問題
1:for Windows中默認拉起的鏡像都在C盤。會致使C盤愈來愈大,建議遷移
若是遷移的盤。好比我這個E盤。路徑中已經存在MobyLinuxVM.vhdx 。是遷移不過的。要刪除,但以前的鏡像都沒有了
若是你想保存,先重命名MobyLinuxVM.vhdx,遷移後。刪除以後的。以前的重命名回來便可
2:共享盤。爲了數據卷掛載用
3:配置鏡像加速(https://hlef8lmt.mirror.aliyuncs.com)
而後能夠去hub.docker上尋找須要的鏡像,官方的mysql有2個鏡像
固然你經過命令也能夠收索到: docker search mysql
首先來看docker mysql
準備須要掛載的目錄和文件,上面我設置的共享盤是D盤,因此掛載的在D盤
my.cnf配置文件,主要是設置mysql的參數
[mysqld] user=mysql character-set-server=utf8 [client] default-character-set=utf8 [mysql] default-character-set=utf8
data是空的。當run的時候,mysql會寫入文件
sql是須要在運行myslq後執行的初始化文件,好比我這裏是給剛建立的用戶名分配權限
這裏爲了說明sql是執行成功的。我在加條。建立數據庫的sql,建立數據庫 docker和user表,並插入一條數據
GRANT ALL PRIVILEGES ON *.* TO 'test'@'%' WITH GRANT OPTION;
Create DATABASE docker;
USE docker;
CREATE TABLE user (ID int auto_increment primary key,name nvarchar(20),address nvarchar(50));
insert into user(name,address)values('劉德華','香港');
初始化後就執行的好處是。不用在run後,去手動執行,關於run後手動執行,
能夠查看我以前的docker安裝mysql http://www.javashuo.com/article/p-edswdmfv-ey.html
所有配置完成後,開始敲命令,如下命令須要去掉註釋
docker run -d -p 3306:3306 --restart always #老是自動重啓。好比系統重啓,該容器會自動啓動 -e MYSQL_USER=test #建立用戶名test -e MYSQL_PASSWORD=123456 #test密碼 -e MYSQL_PASSWORD_HOST=% #test 開啓外部登錄 -e MYSQL_ROOT_PASSWORD=123456 #root密碼 -e MYSQL_ROOT_HOST=% #root開啓外部登錄 -v /d/docker/mysql/my.cnf:/etc/my.cnf #配置文件 -v /d/docker/mysql/sql:/docker-entrypoint-initdb.d #初始化的sql -v /d/docker/mysql/data:/var/lib/mysql #data文件 --name mysql #鏡像名稱 mysql #基於那個鏡像建立容器
執行成功沒有異常後。經過 docker ps 能夠查看運行的容器,若是沒有, 那就經過 docker ps -a 必定會有的
如今能夠經過Navicat鏈接試試
建立了docker庫。user表也有數據,能看到mysql庫,說明test用戶是有權限的
當我使用mysql-server 鏡像時,建立容器會沒法啓動
能夠看到。啓動失敗後。又繼續重啓,由於參數指定了restart always
輸入命令 docker logs mysql 查看啓動日誌
最後在my.cnf中加這個,經測試,啓動成功,就不一一放圖了
數據庫準備好了,那麼就快速的構建一個net core api 接口
1:引入NugGet包,MySql.Data.EntityFrameworkCore
2:建立DbContext
using Docker.Api.Model; using Microsoft.EntityFrameworkCore; using System; using System.Collections.Generic; using System.Linq; using System.Threading.Tasks; namespace Docker.Api.Data { public class DbUserInfoContext : DbContext { public DbUserInfoContext(DbContextOptions<DbUserInfoContext> options) : base(options) { } public DbSet<UseInfo> userInfos { get; set; } /// <summary> /// 模型建立時觸發 /// </summary> /// <param name="modelBuilder"></param> protected override void OnModelCreating(ModelBuilder modelBuilder) { /* 修改表名和主鍵,user對應數據庫的表,mysql默認是區分大小寫的 查看:show variables like '%lower%'; lower_case_table_names 爲 0 區分,1 不區分 */ modelBuilder.Entity<UseInfo>(b => b.ToTable("user").HasKey(u => u.id)); //or //modelBuilder.Entity<user>() // .ToTable("user") // .HasKey(u => u.id); base.OnModelCreating(modelBuilder); } } }
3:添加UserInfo控制器
1 using System; 2 using System.Collections.Generic; 3 using System.Linq; 4 using System.Threading.Tasks; 5 using Docker.Api.Data; 6 using Microsoft.AspNetCore.Http; 7 using Microsoft.AspNetCore.Mvc; 8 using Microsoft.EntityFrameworkCore; 9 10 namespace Docker.Api.Controllers 11 { 12 [Route("api/[controller]")] 13 [ApiController] 14 public class UserInfoController : ControllerBase 15 { 16 private DbUserInfoContext _DbUserInfoContext; 17 public UserInfoController(DbUserInfoContext context) 18 { 19 _DbUserInfoContext = context; 20 } 21 [HttpGet] 22 public async Task<IActionResult> Get() 23 { 24 return new JsonResult(await _DbUserInfoContext.userInfos.FirstOrDefaultAsync()); 25 } 26 } 27 }
4:配置sql鏈接字符串: server=localhost;port=3306;userid=test;password=123456;database=docker
run項目。訪問能成功獲取信息
接下來咱們把這個api也打包成鏡像,而後基於該鏡像建立容器,而後鏈接mysql鏡像的容器。這稱之爲容器互連
容器互連有3種方式
1:Link方式。已經被docker淘汰,docker官方不推薦使用該方式
2:Bridger,橋接的方式,單臺機器用
3:Overlay 適用於集羣時候用
Overlay 就我目前環境不適合測試,集羣也不懂。就不搞了
說說LInk和Bridger方式,具體理論知識請看docker官方文檔。我這裏只實踐
如今一切來回憶下
剛上面打包控制檯應用程序用的是:microsoft/dotnet 鏡像
而後後面帶上tag
好比:
Micirosoft/dotnet:sdk |
包含了運行時和sdk命令,打包後會很大,由於包含sdk,通常用於測試環境 |
Microsoft/dotnet:<version>-runtime |
包含運行時,不包含sdk,打包後就很小了,通常用於正式環境 |
Microsoft/dotnet:<version>-runtime-deps |
打包的時候,會自包含runtime,也就是部署的機器有沒有runtime是沒有關係 上面2種,必須機器要包含core環境 |
修改程序port運行在80上
編寫api的Dockerfile
我這裏用的sdk,由於要用到sdk命令好比dotnet restore,dotnet publish
若是已經publish的文件,直接用runtime會方便不少。上面也有說起
1 #FROM mcr.microsoft.com/dotnet/core/sdk:2.2 AS build 2 FROM microsoft/dotnet:2.2-sdk AS build 3 WORKDIR /src 4 WORKDIR /source 5 #這裏的後面的 . 就是/source 路徑 6 #或者 COPY *.csproj /source 7 COPY *.csproj . 8 RUN dotnet restore 9 COPY . . 10 # 發佈到 /source/out 下 11 RUN dotnet publish -c Release -o out 12 13 #FROM mcr.microsoft.com/dotnet/core/runtime:2.2 14 FROM microsoft/dotnet:2.2-aspnetcore-runtime 15 WORKDIR /app 16 COPY --from=build /source/out . 17 EXPOSE 80 18 ENTRYPOINT ["dotnet","Docker.Api.dll"]
開始build項目 docker build -t cn/myapi .
能夠看到。這裏沒有指定tag。因此默認是latest,size也不大
成功後開始run一個容器,不過這以前要先:
準備掛載目錄。由於配置文件 appsettings 會須要動態配置,因此掛載出來
還有,好比一個網站都有log日誌,這些也須要掛載出來。便於管理。
我這裏就只掛載appsettings.json
執行命令:
docker run -d -p 80:80 --restart always --link mysql:mysqldb -v /d/docker/myapi/appsettings.json:/app/appsettings.json --name api cn/myapi
分析:
--restart always :老是重啓
-d:是在後臺執行
-p 80:80 :第一個80是暴露給外部的。第二個80是程序的。
--link mysql:mysqldb : mysql是容器名稱,mysqldb是自定義名稱,能夠理解爲服務器
-v /d/docker/myapi/appsettings.json:/app/appsettings.json:這裏就是掛載外部的數據捲了
也許你會問。我怎麼知道這個路徑的:/app/appsettings.json。從編寫的dockerfile能分析出來,待會也能夠進入容器看看
最有的工做目錄是 根路徑下: /app
而後經過頁面訪問試試
發現依然沒法訪問,由於修改appsettings.json的鏈接方式
記住這裏是修改D:\docker\myapi\appsettings.json ,由於已經掛載出來
把server改爲mysqldb,而後重啓容器: docker restart api
再次刷新頁面
咱們 進入容器看看: docker exec -it api bash 能夠看到根目錄下存在app目錄
進入app目錄
我的認爲link方式是最簡單的。在這3種中,接下來看看Bridge方式
1:首先建立一個網絡 network,名稱叫api2bridge
docker network create -d bridge api2bridge
經過: docker network ls 能夠查看到已經建立成功
2:實例化容器
爲了區別於上面的80端口,這裏新增一個8081
docker run -d -p 8081:80 --restart always -v /d/docker/myapi/appsettings.json:/app/appsettings.json --net api2bridge --name api2 cn/myapi
建立容器的時候,自定network 這裏的--net api2bridge 就是上面的bridge
3:鏈接2個容器,經過: docker network connect api2bridge mysql 把api2和mysql鏈接起來
4:修改appsettings.json server=mysql
5 : restart 容器,若是是在建立容器前修改的配置文件。是不須要重啓的,測試經過
看看這兩個容器是怎麼鏈接的。經過命令: docker inspect api2bridge 能夠查看對象的元數據(容器或者網絡)
分別看看;
docker inspect api2
docker inspect mysql
你會發現mysql有個"IPAddress":地址,
上面咱們在api2中的appsettings.json的server是直接些的容器名稱:mysql。也能夠直接些這個ip地址。好比: server=172.20.0.3 一樣是能夠的。
Overlay方式就不講了。由於我也不知道。哈哈
經過這幾個例子你會發現。2個容器要部署2個,若是項目依賴mysql,redis,MQ等等。那得部署屢次,如此重複性的工做會影響效率
因此有了docker-compose,compose
參考:https://yeasy.gitbooks.io/docker_practice/content/compose/install.html
安裝:
sudo curl -L https://github.com/docker/compose/releases/download/1.17.1/docker-compose-`uname -s`-`uname -m` > /usr/local/bin/docker-compose sudo chmod +x /usr/local/bin/docker-compose
安裝完成後,能夠經過: docker-compose --version 查看版本
經過: docker-compose --help 查看基本的命令
不過我英文很差,就經過百度翻譯了,翻譯得有點生硬。僅供參考
Commands:
build 創建或重建服務
bundle 從撰寫文件生成Docker捆綁包
config 驗證並查看撰寫文件
create 建立服務
down 中止並刪除容器、網絡、圖像和卷
events 從容器接收實時事件
exec 在正在運行的容器中執行命令
help 獲取有關命令的幫助
images 列表圖像
kill 殺死容器
logs 查看容器的輸出
pause 暫停服務
port 打印端口綁定的公共端口
ps 列表容器
pull 拉取服務圖像
push 推送服務圖像
restart 從新啓動服務
rm 移除中止的容器
run 運行一次性命令
scale 設置服務的容器數
start 啓動服務
stop 中止服務
top 顯示正在運行的進程
unpause 取消暫停服務
up 建立和啓動容器
version 顯示Docker撰寫版本信息
目前爲止已經有3個容器了,
爲了區別於以前的mysql和api和api2,這裏命名要修改,編寫在程序根目錄下添加docker-compose.yml文件
compose用的是yml語法。能夠參考阮一峯些的文章
http://www.ruanyifeng.com/blog/2016/07/yaml.html
項目準備。依然在上面的api項目中添磚加瓦
還記得上面初始化的建立docker庫,user表嗎。這裏咱們經過在代碼中來實現,
場景:建立myslq的時候,判斷數據庫是否有數據,不然新增一條數據
技術棧:項目依賴mysql,redis,其實我工做中用的都是mssql,因此待會也會介紹
1:init.sql 只保留一條sql語句
2:新增UserInit類。用於初始化數據
using Microsoft.AspNetCore.Builder; using Microsoft.EntityFrameworkCore; using Microsoft.Extensions.DependencyInjection; using Microsoft.Extensions.Logging; using System; using System.Collections.Generic; using System.Linq; using System.Threading.Tasks; namespace Docker.Api.Data { public class UserInit { private ILogger<UserInit> _logger; public UserInit(ILogger<UserInit> logger) { _logger = logger; } public static async Task InitData(IApplicationBuilder app, ILoggerFactory loggerFactory) { using (var scope = app.ApplicationServices.CreateScope()) { var context = scope.ServiceProvider.GetService<DbUserInfoContext>(); var logger = scope.ServiceProvider.GetService<ILogger<UserInit>>(); logger.LogDebug("begin mysql init"); context.Database.Migrate(); if (context.userInfos.Count() <= 0) { context.userInfos.Add(new Model.UseInfo { name = "admin", address = "博客園" }); context.SaveChanges(); } } await Task.CompletedTask; } } }
程序啓動調用:
3:實體類
using System; using System.Collections.Generic; using System.Linq; using System.Threading.Tasks; namespace Docker.Api.Model { public class UseInfo { public int id { get; set; } public string name { get; set; } public string address { get; set; } } }
4:DbContext 上面也列出,這裏就不展現了
5:RedisHelper網絡有。這裏也不提了
只准備2個接口。用於測試redis。一個讀,一個寫
using System; using System.Collections.Generic; using System.Linq; using System.Threading.Tasks; using Microsoft.AspNetCore.Http; using Microsoft.AspNetCore.Mvc; namespace Docker.Api.Controllers { [Route("api/[controller]")] [ApiController] public class RedisController : ControllerBase { [HttpPost] public void Post() { RedisCommon.GetRedis().StringSet("docker", "hello", TimeSpan.FromMinutes(1)); } [HttpGet] public string Get() { var docker = RedisCommon.GetRedis().GetStringKey("docker"); if (docker.HasValue) return docker.ToString(); return "empty"; } } }
6:根據Model生成Migration,這裏簡單過一下,具體參考我以前的:http://www.javashuo.com/article/p-scqvowfx-gk.html
調出程序包管理控制檯
輸入: Add-Migration init
若是成功了就會這樣:
編寫docker-compose.yml 文件,我這裏的註釋是便於理解。儘可能不要寫
注:我是直接在項目中建立的文本文件,而後修改後綴名
在網絡上看到說。若是是在外部建立的記事本。要修改編碼爲:ASCII編碼格式,我未測試
version: '3' services: db: image: mysql container_name: 'mysql01' command: --character-set-server=utf8 --collation-server=utf8_general_ci restart: always ports: - '3307:3306' environment: MYSQL_USER: test MYSQL_PASSWORD: 123456 MYSQL_PASSWORD_HOST: '%' MYSQL_ROOT_PASSWORD: 123456 MYSQL_ROOT_HOST: '%' volumes: - /d/docker/mysql02/my.cnf:/etc/my.cnf - /d/docker/mysql02/data:/var/lib/mysql - /d/docker/mysql02/SqlInit:/docker-entrypoint-initdb.d redis: image: redis container_name: 'redis' command: redis-server /usr/local/etc/redis/redis.conf restart: always ports: - '6379:6379' environment: requirepass: 123456 #redis密碼 appendonly: 'yes' #redis是否持久化 volumes: - /d/docker/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf - /d/docker/redis/data:/data #這裏會保存持久化數據 web: build: . #會執行當前目錄下面的dockerfile文件 container_name: 'api3' #容器名稱 restart: always # web依賴於db,若是web比db啓動快。就鏈接不上db致使web異常,web容器啓動失敗,restart能夠不斷重試,直到鏈接爲止 volumes: - /d/docker/myapi/appsettings.json:/app/appsettings.json ports: - '8082:80' depends_on: #依賴db容器,並不表明執行順序 - db - redis
若是想看編寫的yml文件是否正確,能夠去在線的網站,驗證是否正確,好比:http://nodeca.github.io/js-yaml/
切換到當前目錄輸入: docker-compose build 會開始build項目成鏡像
查看鏡像:名字叫dockerapi_web
輸入命令: docker-compose up 會開始建立容器並啓動
輸出的日誌太多。這裏只點幾個有用的看
EFcore插入Migration歷史記錄
建立表:
直到最後,程序阻塞。顯示成功,由於這裏沒用用 -d 會阻塞,調試的時候不建議 -d
而後新打開一個PowerShell,輸入docker ps 查看運行的容器
分別測試是否成功
一樣驗證redis,用RedisDesktopManager鏈接
從容器能夠看出api3端口是8082,嘗試訪問下
測試寫redis,打開Postman寫入Redis
寫人成功
那麼讀取就不是什麼大問題了
問題彙總:
若是你修改了代碼,須要從新build。那麼先刪除容器: docker-compose down 會中止容器並刪除
docker-compose ps 查看容器列表
docker-compose up -d 後端運行,不阻塞前端
docker-compose restart 重啓全部容器。
自此全部容器成功運行,但我感受還不夠,由於一直都是在windos上玩。而沒有上CentOS7,可我又不缺CentOS環境。因此要玩一把
技術棧:Jexus,mysql,mssql,redis
關於jexus部署net core 能夠參考我前面寫的文章:http://www.javashuo.com/article/p-wlxlvbyq-m.html
既然要加入新的成員。jexus 和 mssql,那麼就得修改docker-compose文件
在經過docker-compose統一打包前,咱們先來單獨玩玩mssql
準備數據卷掛載目錄
data:保存數據庫文件
sql:執行的腳本。mssql沒有mysql的docker-entrypoint-initdb.d 掛載,啓動mysql就執行sql。這裏sql文件夾
雖然保存的是.sql文件。但要手動執行,不知道是否是我沒有找到具體的方案
sql裏面放一個init.sql文件。編寫sql腳本以下
這裏要注意一點,一條語句完成必需要帶一個Go語句
參考官方文檔:
鏡像文檔:https://hub.docker.com/_/microsoft-mssql-server
//註釋部分 docker run -d -p 1433:1433 \ -e ACCEPT_EULA=Y \ #確認您接受最終用戶許可協議。 -e SA_PASSWORD=DockerPwd123 \ #強大的系統管理員(SA)密碼:至少8個字符,包括大寫,小寫字母,基數爲10的數字和/或非字母數字符號。 -e MSSQL_PID=Express \ #版本(Developer,Express,Enterprise,EnterpriseCore)默認值:Developer -v /docker/mssql:/var/opt/mssql \ # 映射數據庫 v /d/docker/mssql/sql:/script #把須要執行的腳本放這裏,script路徑隨便改,不是初始化執行,是手到執行 --name mssql #容器名稱 mcr.microsoft.com/mssql/server #鏡像
執行成功後。數據卷掛載目錄。生成了文件
此時data也有默認的數據庫了
經過 MSSMS( Microsoft SQL Server Management Studio )鏈接試試
剛上面說了sql中文件是沒有被執行的。必須手動執行。
手動執行前,先來看看其餘一些相關命令
進入容器後: docker exec -it mssql bash
登錄數據庫:localhost也能夠用指定的ip代替,若是有端口。則帶端口號便可
/opt/mssql-tools/bin/sqlcmd -S localhost -U SA -P DockerPwd123 -S 是服務器,無論端口是多少,都不用寫 -U 是用戶名 -P 是密碼
若是出現 1> 說明的登錄成功了
能夠輸入語句:select getdate() 試試,回車後,須要加Go語句,不過日期怎麼不對?好像是相差8個時區
修改時區,能夠經過TZ變量
docker run -e TZ="Asia/Shanghai" -e 'ACCEPT_EULA=Y' -e 'MSSQL_SA_PASSWORD=DockerPwd123' -e MSSQL_PID=Express -p 1433:1433 --name mssql -d mcr.microsoft.com/mssql/server
官網文檔:https://docs.microsoft.com/zh-cn/sql/linux/sql-server-linux-configure-docker?view=sql-server-2017#tz
執行sql中的文件
登錄容器後執行操做: /opt/mssql-tools/bin/sqlcmd -S localhost -U SA -P DockerPwd123 -i /script/init.sql
掛載目錄也有,這樣就算容器沒法進入。數據庫也存在
因爲時間問題,docker-compose 就不加入mssql,只加jexus,修改docker-compose以下
直接上docker-compose文件
version: '3' services: db: image: mysql container_name: 'mysql' command: --character-set-server=utf8 --collation-server=utf8_general_ci restart: always ports: - '3306:3306' environment: MYSQL_USER: test MYSQL_PASSWORD: 123456 MYSQL_PASSWORD_HOST: '%' MYSQL_ROOT_PASSWORD: 123456 MYSQL_ROOT_HOST: '%' volumes: - /docker/mysq/my.cnf:/etc/my.cnf - /docker/mysql/data:/var/lib/mysql - /docker/mysql/sql:/docker-entrypoint-initdb.d redis: image: redis container_name: 'redis' command: redis-server /usr/local/etc/redis/redis.conf restart: always ports: - '6379:6379' environment: requirepass: 123456 #redis密碼 appendonly: 'yes' #redis是否持久化 volumes: - /docker/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf - /docker/redis/data:/data #這裏會保存持久化數據 jexus: image: byniqing/jexus container_name: 'jexus' ports: - '80:8803'#前一個80是暴露外部的,後一個8803是jexus監聽的的。也就是配置文件中port=8803,此端口必需要開通 restart: always volumes: - /docker/jexus/www:/var/www - /docker/jexus/siteconf:/usr/jexus/siteconf - /docker/jexus/log:/usr/jexus/log #depends_on: #- web web: build: . #會執行當前目錄下面的dockerfile文件 container_name: 'api' #容器名稱 restart: always # web依賴於db,若是web比db啓動快。就鏈接不上db致使web異常,web容器啓動失敗,restart能夠不斷重試,直到鏈接爲止 volumes: - /docker/myapi/appsettings.json:/app/appsettings.json ports: - '8802:80'#8802端口必須開啓,若是是阿里雲。添加入棧規則,jexus配置。代理到8802便可,瀏覽器直接訪問這個port也能訪問 depends_on: #依賴db容器,並不表明執行順序 - db - redis
分析:
1:jexus我用了本身打包後的鏡像。而後push到hub.docker上去了,你們也能夠打包一個本身的。方便之後測試,升級,固然直接用給我這個也能夠,dockerfile 以下:
FROM debian:latest MAINTAINER xxx <xxx@xxx.com> RUN apt-get update \ && apt-get -y install wget \ && cd /usr \ && wget https://www.linuxdot.net/down/jexus-5.8.3-x64.tar.gz \ && tar -zxvf jexus-5.8.3-x64.tar.gz \ && apt-get -y autoremove --purge wget \ && rm -rf /var/lib/apt/lists/* jexus-5.8.3-x64.tar.gz EXPOSE 80 WORKDIR /usr/jexus CMD /usr/jexus/jwss
2:web中的ports暴露了8802端口,須要開通該端口
3:jexus文件配置:reproxy=/ ip:8802(繞過容器內部,直接訪問容器,因此直接在瀏覽器也是直接能夠訪問的),或者reproxy=/ web:80 (容器互聯)
而後你會發現經過ip:80和ip:8802都能訪問,
一個是經過jexus訪問。一個是繞過了jexus直接訪問了api接口,那怎麼行呢,那麼jexus就沒有存在的用意了
因此:咱們應該不暴露web端口,即屏蔽掉ports
而後從新修改jexus,經過容器名稱訪問: reproxy=/ web:80
由於compose打包後後是在同一個網絡中:
好比: docker-compose ps 能夠查看當前服務下的全部容器
那怎麼判斷當前容器使用的那些網絡呢?
能夠查看當前某個容器的元數據:好比咱們來看mysql的: docker inspect mysql 也能夠看出Networks節點信息
經過: docker network ls 查看當前網絡,存在 dockerapi_default
那麼就能夠經過: docker inspect dockerapi_default 查看下dockerapi_default網絡的元數據
能夠看到Containers節點下。docker-compose中定義的4個容器,而且ip4都在192.168.48.x/20
不知道你跟是否有同一個疑問,雖然api沒有暴露接口,但api和jexus。tcp都是80
那我直接訪問ip地址怎麼肯定必定就是訪問的jexus。而不是Api呢。那麼咱們來改造一下,jexus暴露8802,api內部依然是80
重新打包,查看容器:
而後瀏覽器輸入:ip:8802/api/values 訪問成功
輸入:ip/aip/values 訪問失敗
那麼這樣就達到了只能經過jexus訪問到個人aip了
使用數據卷,文件掛載到宿主機就是爲了方便修改,這裏拿redis爲例
咱們在建立redis的時候有個掛載目錄爲: /docker/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf
redis.conf這個就是redis的配置文件。能夠自行修改,
去網上下載一個對應版本的配置文件放進去便可:http://download.redis.io/releases/
好比上面的默認密碼是:123456,咱們來改爲7890
1:註釋掉,bind 127.0.0.1
2:修改,requirepass 值,是否是發現requirepass 跟docker-compose 中變量是同一個
3:重啓redis
4:測試鏈接成功,
進入redis容器。查看 /usr/local/etc/redis/redis.conf
你會發現,redis.conf是同步更新的,這裏就不截圖了
放一個redis.confi文件
1 # Redis configuration file example 2 3 # Note on units: when memory size is needed, it is possible to specify 4 # it in the usual form of 1k 5GB 4M and so forth: 5 # 6 # 1k => 1000 bytes 7 # 1kb => 1024 bytes 8 # 1m => 1000000 bytes 9 # 1mb => 1024*1024 bytes 10 # 1g => 1000000000 bytes 11 # 1gb => 1024*1024*1024 bytes 12 # 13 # units are case insensitive so 1GB 1Gb 1gB are all the same. 14 15 ################################## INCLUDES ################################### 16 17 # Include one or more other config files here. This is useful if you 18 # have a standard template that goes to all Redis servers but also need 19 # to customize a few per-server settings. Include files can include 20 # other files, so use this wisely. 21 # 22 # Notice option "include" won't be rewritten by command "CONFIG REWRITE" 23 # from admin or Redis Sentinel. Since Redis always uses the last processed 24 # line as value of a configuration directive, you'd better put includes 25 # at the beginning of this file to avoid overwriting config change at runtime. 26 # 27 # If instead you are interested in using includes to override configuration 28 # options, it is better to use include as the last line. 29 # 30 # include .\path\to\local.conf 31 # include c:\path\to\other.conf 32 33 ################################## NETWORK ##################################### 34 35 # By default, if no "bind" configuration directive is specified, Redis listens 36 # for connections from all the network interfaces available on the server. 37 # It is possible to listen to just one or multiple selected interfaces using 38 # the "bind" configuration directive, followed by one or more IP addresses. 39 # 40 # Examples: 41 # 42 # bind 192.168.1.100 10.0.0.1 43 # bind 127.0.0.1 ::1 44 # 45 # ~~~ WARNING ~~~ If the computer running Redis is directly exposed to the 46 # internet, binding to all the interfaces is dangerous and will expose the 47 # instance to everybody on the internet. So by default we uncomment the 48 # following bind directive, that will force Redis to listen only into 49 # the IPv4 lookback interface address (this means Redis will be able to 50 # accept connections only from clients running into the same computer it 51 # is running). 52 # 53 # IF YOU ARE SURE YOU WANT YOUR INSTANCE TO LISTEN TO ALL THE INTERFACES 54 # JUST COMMENT THE FOLLOWING LINE. 55 # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 56 #bind 127.0.0.1 57 58 # Protected mode is a layer of security protection, in order to avoid that 59 # Redis instances left open on the internet are accessed and exploited. 60 # 61 # When protected mode is on and if: 62 # 63 # 1) The server is not binding explicitly to a set of addresses using the 64 # "bind" directive. 65 # 2) No password is configured. 66 # 67 # The server only accepts connections from clients connecting from the 68 # IPv4 and IPv6 loopback addresses 127.0.0.1 and ::1, and from Unix domain 69 # sockets. 70 # 71 # By default protected mode is enabled. You should disable it only if 72 # you are sure you want clients from other hosts to connect to Redis 73 # even if no authentication is configured, nor a specific set of interfaces 74 # are explicitly listed using the "bind" directive. 75 protected-mode yes 76 77 # Accept connections on the specified port, default is 6379 (IANA #815344). 78 # If port 0 is specified Redis will not listen on a TCP socket. 79 port 6379 80 81 # TCP listen() backlog. 82 # 83 # In high requests-per-second environments you need an high backlog in order 84 # to avoid slow clients connections issues. Note that the Linux kernel 85 # will silently truncate it to the value of /proc/sys/net/core/somaxconn so 86 # make sure to raise both the value of somaxconn and tcp_max_syn_backlog 87 # in order to get the desired effect. 88 tcp-backlog 511 89 90 # Unix socket. 91 # 92 # Specify the path for the Unix socket that will be used to listen for 93 # incoming connections. There is no default, so Redis will not listen 94 # on a unix socket when not specified. 95 # 96 # unixsocket /tmp/redis.sock 97 # unixsocketperm 700 98 99 # Close the connection after a client is idle for N seconds (0 to disable) 100 timeout 0 101 102 # TCP keepalive. 103 # 104 # If non-zero, use SO_KEEPALIVE to send TCP ACKs to clients in absence 105 # of communication. This is useful for two reasons: 106 # 107 # 1) Detect dead peers. 108 # 2) Take the connection alive from the point of view of network 109 # equipment in the middle. 110 # 111 # On Linux, the specified value (in seconds) is the period used to send ACKs. 112 # Note that to close the connection the double of the time is needed. 113 # On other kernels the period depends on the kernel configuration. 114 # 115 # A reasonable value for this option is 60 seconds. 116 tcp-keepalive 0 117 118 ################################# GENERAL ##################################### 119 120 # By default Redis does not run as a daemon. Use 'yes' if you need it. 121 # Note that Redis will write a pid file in /var/run/redis.pid when daemonized. 122 # NOT SUPPORTED ON WINDOWS daemonize no 123 124 # If you run Redis from upstart or systemd, Redis can interact with your 125 # supervision tree. Options: 126 # supervised no - no supervision interaction 127 # supervised upstart - signal upstart by putting Redis into SIGSTOP mode 128 # supervised systemd - signal systemd by writing READY=1 to $NOTIFY_SOCKET 129 # supervised auto - detect upstart or systemd method based on 130 # UPSTART_JOB or NOTIFY_SOCKET environment variables 131 # Note: these supervision methods only signal "process is ready." 132 # They do not enable continuous liveness pings back to your supervisor. 133 # NOT SUPPORTED ON WINDOWS supervised no 134 135 # If a pid file is specified, Redis writes it where specified at startup 136 # and removes it at exit. 137 # 138 # When the server runs non daemonized, no pid file is created if none is 139 # specified in the configuration. When the server is daemonized, the pid file 140 # is used even if not specified, defaulting to "/var/run/redis.pid". 141 # 142 # Creating a pid file is best effort: if Redis is not able to create it 143 # nothing bad happens, the server will start and run normally. 144 # NOT SUPPORTED ON WINDOWS pidfile /var/run/redis.pid 145 146 # Specify the server verbosity level. 147 # This can be one of: 148 # debug (a lot of information, useful for development/testing) 149 # verbose (many rarely useful info, but not a mess like the debug level) 150 # notice (moderately verbose, what you want in production probably) 151 # warning (only very important / critical messages are logged) 152 loglevel notice 153 154 # Specify the log file name. Also 'stdout' can be used to force 155 # Redis to log on the standard output. 156 logfile "" 157 158 # To enable logging to the Windows EventLog, just set 'syslog-enabled' to 159 # yes, and optionally update the other syslog parameters to suit your needs. 160 # If Redis is installed and launched as a Windows Service, this will 161 # automatically be enabled. 162 # syslog-enabled no 163 164 # Specify the source name of the events in the Windows Application log. 165 # syslog-ident redis 166 167 # Set the number of databases. The default database is DB 0, you can select 168 # a different one on a per-connection basis using SELECT <dbid> where 169 # dbid is a number between 0 and 'databases'-1 170 databases 16 171 172 ################################ SNAPSHOTTING ################################ 173 # 174 # Save the DB on disk: 175 # 176 # save <seconds> <changes> 177 # 178 # Will save the DB if both the given number of seconds and the given 179 # number of write operations against the DB occurred. 180 # 181 # In the example below the behaviour will be to save: 182 # after 900 sec (15 min) if at least 1 key changed 183 # after 300 sec (5 min) if at least 10 keys changed 184 # after 60 sec if at least 10000 keys changed 185 # 186 # Note: you can disable saving completely by commenting out all "save" lines. 187 # 188 # It is also possible to remove all the previously configured save 189 # points by adding a save directive with a single empty string argument 190 # like in the following example: 191 # 192 # save "" 193 194 save 900 1 195 save 300 10 196 save 60 10000 197 198 # By default Redis will stop accepting writes if RDB snapshots are enabled 199 # (at least one save point) and the latest background save failed. 200 # This will make the user aware (in a hard way) that data is not persisting 201 # on disk properly, otherwise chances are that no one will notice and some 202 # disaster will happen. 203 # 204 # If the background saving process will start working again Redis will 205 # automatically allow writes again. 206 # 207 # However if you have setup your proper monitoring of the Redis server 208 # and persistence, you may want to disable this feature so that Redis will 209 # continue to work as usual even if there are problems with disk, 210 # permissions, and so forth. 211 stop-writes-on-bgsave-error yes 212 213 # Compress string objects using LZF when dump .rdb databases? 214 # For default that's set to 'yes' as it's almost always a win. 215 # If you want to save some CPU in the saving child set it to 'no' but 216 # the dataset will likely be bigger if you have compressible values or keys. 217 rdbcompression yes 218 219 # Since version 5 of RDB a CRC64 checksum is placed at the end of the file. 220 # This makes the format more resistant to corruption but there is a performance 221 # hit to pay (around 10%) when saving and loading RDB files, so you can disable it 222 # for maximum performances. 223 # 224 # RDB files created with checksum disabled have a checksum of zero that will 225 # tell the loading code to skip the check. 226 rdbchecksum yes 227 228 # The filename where to dump the DB 229 dbfilename dump.rdb 230 231 # The working directory. 232 # 233 # The DB will be written inside this directory, with the filename specified 234 # above using the 'dbfilename' configuration directive. 235 # 236 # The Append Only File will also be created inside this directory. 237 # 238 # Note that you must specify a directory here, not a file name. 239 dir ./ 240 241 ################################# REPLICATION ################################# 242 243 # Master-Slave replication. Use slaveof to make a Redis instance a copy of 244 # another Redis server. A few things to understand ASAP about Redis replication. 245 # 246 # 1) Redis replication is asynchronous, but you can configure a master to 247 # stop accepting writes if it appears to be not connected with at least 248 # a given number of slaves. 249 # 2) Redis slaves are able to perform a partial resynchronization with the 250 # master if the replication link is lost for a relatively small amount of 251 # time. You may want to configure the replication backlog size (see the next 252 # sections of this file) with a sensible value depending on your needs. 253 # 3) Replication is automatic and does not need user intervention. After a 254 # network partition slaves automatically try to reconnect to masters 255 # and resynchronize with them. 256 # 257 # slaveof <masterip> <masterport> 258 259 # If the master is password protected (using the "requirepass" configuration 260 # directive below) it is possible to tell the slave to authenticate before 261 # starting the replication synchronization process, otherwise the master will 262 # refuse the slave request. 263 # 264 # masterauth <master-password> 265 266 # When a slave loses its connection with the master, or when the replication 267 # is still in progress, the slave can act in two different ways: 268 # 269 # 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will 270 # still reply to client requests, possibly with out of date data, or the 271 # data set may just be empty if this is the first synchronization. 272 # 273 # 2) if slave-serve-stale-data is set to 'no' the slave will reply with 274 # an error "SYNC with master in progress" to all the kind of commands 275 # but to INFO and SLAVEOF. 276 # 277 slave-serve-stale-data yes 278 279 # You can configure a slave instance to accept writes or not. Writing against 280 # a slave instance may be useful to store some ephemeral data (because data 281 # written on a slave will be easily deleted after resync with the master) but 282 # may also cause problems if clients are writing to it because of a 283 # misconfiguration. 284 # 285 # Since Redis 2.6 by default slaves are read-only. 286 # 287 # Note: read only slaves are not designed to be exposed to untrusted clients 288 # on the internet. It's just a protection layer against misuse of the instance. 289 # Still a read only slave exports by default all the administrative commands 290 # such as CONFIG, DEBUG, and so forth. To a limited extent you can improve 291 # security of read only slaves using 'rename-command' to shadow all the 292 # administrative / dangerous commands. 293 slave-read-only yes 294 295 # Replication SYNC strategy: disk or socket. 296 # 297 # ------------------------------------------------------- 298 # WARNING: DISKLESS REPLICATION IS EXPERIMENTAL CURRENTLY 299 # ------------------------------------------------------- 300 # 301 # New slaves and reconnecting slaves that are not able to continue the replication 302 # process just receiving differences, need to do what is called a "full 303 # synchronization". An RDB file is transmitted from the master to the slaves. 304 # The transmission can happen in two different ways: 305 # 306 # 1) Disk-backed: The Redis master creates a new process that writes the RDB 307 # file on disk. Later the file is transferred by the parent 308 # process to the slaves incrementally. 309 # 2) Diskless: The Redis master creates a new process that directly writes the 310 # RDB file to slave sockets, without touching the disk at all. 311 # 312 # With disk-backed replication, while the RDB file is generated, more slaves 313 # can be queued and served with the RDB file as soon as the current child producing 314 # the RDB file finishes its work. With diskless replication instead once 315 # the transfer starts, new slaves arriving will be queued and a new transfer 316 # will start when the current one terminates. 317 # 318 # When diskless replication is used, the master waits a configurable amount of 319 # time (in seconds) before starting the transfer in the hope that multiple slaves 320 # will arrive and the transfer can be parallelized. 321 # 322 # With slow disks and fast (large bandwidth) networks, diskless replication 323 # works better. 324 repl-diskless-sync no 325 326 # When diskless replication is enabled, it is possible to configure the delay 327 # the server waits in order to spawn the child that transfers the RDB via socket 328 # to the slaves. 329 # 330 # This is important since once the transfer starts, it is not possible to serve 331 # new slaves arriving, that will be queued for the next RDB transfer, so the server 332 # waits a delay in order to let more slaves arrive. 333 # 334 # The delay is specified in seconds, and by default is 5 seconds. To disable 335 # it entirely just set it to 0 seconds and the transfer will start ASAP. 336 repl-diskless-sync-delay 5 337 338 # Slaves send PINGs to server in a predefined interval. It's possible to change 339 # this interval with the repl_ping_slave_period option. The default value is 10 340 # seconds. 341 # 342 # repl-ping-slave-period 10 343 344 # The following option sets the replication timeout for: 345 # 346 # 1) Bulk transfer I/O during SYNC, from the point of view of slave. 347 # 2) Master timeout from the point of view of slaves (data, pings). 348 # 3) Slave timeout from the point of view of masters (REPLCONF ACK pings). 349 # 350 # It is important to make sure that this value is greater than the value 351 # specified for repl-ping-slave-period otherwise a timeout will be detected 352 # every time there is low traffic between the master and the slave. 353 # 354 # repl-timeout 60 355 356 # Disable TCP_NODELAY on the slave socket after SYNC? 357 # 358 # If you select "yes" Redis will use a smaller number of TCP packets and 359 # less bandwidth to send data to slaves. But this can add a delay for 360 # the data to appear on the slave side, up to 40 milliseconds with 361 # Linux kernels using a default configuration. 362 # 363 # If you select "no" the delay for data to appear on the slave side will 364 # be reduced but more bandwidth will be used for replication. 365 # 366 # By default we optimize for low latency, but in very high traffic conditions 367 # or when the master and slaves are many hops away, turning this to "yes" may 368 # be a good idea. 369 repl-disable-tcp-nodelay no 370 371 # Set the replication backlog size. The backlog is a buffer that accumulates 372 # slave data when slaves are disconnected for some time, so that when a slave 373 # wants to reconnect again, often a full resync is not needed, but a partial 374 # resync is enough, just passing the portion of data the slave missed while 375 # disconnected. 376 # 377 # The bigger the replication backlog, the longer the time the slave can be 378 # disconnected and later be able to perform a partial resynchronization. 379 # 380 # The backlog is only allocated once there is at least a slave connected. 381 # 382 # repl-backlog-size 1mb 383 384 # After a master has no longer connected slaves for some time, the backlog 385 # will be freed. The following option configures the amount of seconds that 386 # need to elapse, starting from the time the last slave disconnected, for 387 # the backlog buffer to be freed. 388 # 389 # A value of 0 means to never release the backlog. 390 # 391 # repl-backlog-ttl 3600 392 393 # The slave priority is an integer number published by Redis in the INFO output. 394 # It is used by Redis Sentinel in order to select a slave to promote into a 395 # master if the master is no longer working correctly. 396 # 397 # A slave with a low priority number is considered better for promotion, so 398 # for instance if there are three slaves with priority 10, 100, 25 Sentinel will 399 # pick the one with priority 10, that is the lowest. 400 # 401 # However a special priority of 0 marks the slave as not able to perform the 402 # role of master, so a slave with priority of 0 will never be selected by 403 # Redis Sentinel for promotion. 404 # 405 # By default the priority is 100. 406 slave-priority 100 407 408 # It is possible for a master to stop accepting writes if there are less than 409 # N slaves connected, having a lag less or equal than M seconds. 410 # 411 # The N slaves need to be in "online" state. 412 # 413 # The lag in seconds, that must be <= the specified value, is calculated from 414 # the last ping received from the slave, that is usually sent every second. 415 # 416 # This option does not GUARANTEE that N replicas will accept the write, but 417 # will limit the window of exposure for lost writes in case not enough slaves 418 # are available, to the specified number of seconds. 419 # 420 # For example to require at least 3 slaves with a lag <= 10 seconds use: 421 # 422 # min-slaves-to-write 3 423 # min-slaves-max-lag 10 424 # 425 # Setting one or the other to 0 disables the feature. 426 # 427 # By default min-slaves-to-write is set to 0 (feature disabled) and 428 # min-slaves-max-lag is set to 10. 429 430 ################################## SECURITY ################################### 431 432 # Require clients to issue AUTH <PASSWORD> before processing any other 433 # commands. This might be useful in environments in which you do not trust 434 # others with access to the host running redis-server. 435 # 436 # This should stay commented out for backward compatibility and because most 437 # people do not need auth (e.g. they run their own servers). 438 # 439 # Warning: since Redis is pretty fast an outside user can try up to 440 # 150k passwords per second against a good box. This means that you should 441 # use a very strong password otherwise it will be very easy to break. 442 # 443 # requirepass foobared 444 requirepass 7890 445 # Command renaming. 446 # 447 # It is possible to change the name of dangerous commands in a shared 448 # environment. For instance the CONFIG command may be renamed into something 449 # hard to guess so that it will still be available for internal-use tools 450 # but not available for general clients. 451 # 452 # Example: 453 # 454 # rename-command CONFIG b840fc02d524045429941cc15f59e41cb7be6c52 455 # 456 # It is also possible to completely kill a command by renaming it into 457 # an empty string: 458 # 459 # rename-command CONFIG "" 460 # 461 # Please note that changing the name of commands that are logged into the 462 # AOF file or transmitted to slaves may cause problems. 463 464 ################################### LIMITS #################################### 465 466 # Set the max number of connected clients at the same time. By default 467 # this limit is set to 10000 clients, however if the Redis server is not 468 # able to configure the process file limit to allow for the specified limit 469 # the max number of allowed clients is set to the current file limit 470 # minus 32 (as Redis reserves a few file descriptors for internal uses). 471 # 472 # Once the limit is reached Redis will close all the new connections sending 473 # an error 'max number of clients reached'. 474 # 475 # maxclients 10000 476 477 # If Redis is to be used as an in-memory-only cache without any kind of 478 # persistence, then the fork() mechanism used by the background AOF/RDB 479 # persistence is unnecessary. As an optimization, all persistence can be 480 # turned off in the Windows version of Redis. This will redirect heap 481 # allocations to the system heap allocator, and disable commands that would 482 # otherwise cause fork() operations: BGSAVE and BGREWRITEAOF. 483 # This flag may not be combined with any of the other flags that configure 484 # AOF and RDB operations. 485 # persistence-available [(yes)|no] 486 487 # Don't use more memory than the specified amount of bytes. 488 # When the memory limit is reached Redis will try to remove keys 489 # according to the eviction policy selected (see maxmemory-policy). 490 # 491 # If Redis can't remove keys according to the policy, or if the policy is 492 # set to 'noeviction', Redis will start to reply with errors to commands 493 # that would use more memory, like SET, LPUSH, and so on, and will continue 494 # to reply to read-only commands like GET. 495 # 496 # This option is usually useful when using Redis as an LRU cache, or to set 497 # a hard memory limit for an instance (using the 'noeviction' policy). 498 # 499 # WARNING: If you have slaves attached to an instance with maxmemory on, 500 # the size of the output buffers needed to feed the slaves are subtracted 501 # from the used memory count, so that network problems / resyncs will 502 # not trigger a loop where keys are evicted, and in turn the output 503 # buffer of slaves is full with DELs of keys evicted triggering the deletion 504 # of more keys, and so forth until the database is completely emptied. 505 # 506 # In short... if you have slaves attached it is suggested that you set a lower 507 # limit for maxmemory so that there is some free RAM on the system for slave 508 # output buffers (but this is not needed if the policy is 'noeviction'). 509 # 510 # WARNING: not setting maxmemory will cause Redis to terminate with an 511 # out-of-memory exception if the heap limit is reached. 512 # 513 # NOTE: since Redis uses the system paging file to allocate the heap memory, 514 # the Working Set memory usage showed by the Windows Task Manager or by other 515 # tools such as ProcessExplorer will not always be accurate. For example, right 516 # after a background save of the RDB or the AOF files, the working set value 517 # may drop significantly. In order to check the correct amount of memory used 518 # by the redis-server to store the data, use the INFO client command. The INFO 519 # command shows only the memory used to store the redis data, not the extra 520 # memory used by the Windows process for its own requirements. Th3 extra amount 521 # of memory not reported by the INFO command can be calculated subtracting the 522 # Peak Working Set reported by the Windows Task Manager and the used_memory_peak 523 # reported by the INFO command. 524 # 525 # maxmemory <bytes> 526 527 # MAXMEMORY POLICY: how Redis will select what to remove when maxmemory 528 # is reached. You can select among five behaviors: 529 # 530 # volatile-lru -> remove the key with an expire set using an LRU algorithm 531 # allkeys-lru -> remove any key according to the LRU algorithm 532 # volatile-random -> remove a random key with an expire set 533 # allkeys-random -> remove a random key, any key 534 # volatile-ttl -> remove the key with the nearest expire time (minor TTL) 535 # noeviction -> don't expire at all, just return an error on write operations 536 # 537 # Note: with any of the above policies, Redis will return an error on write 538 # operations, when there are no suitable keys for eviction. 539 # 540 # At the date of writing these commands are: set setnx setex append 541 # incr decr rpush lpush rpushx lpushx linsert lset rpoplpush sadd 542 # sinter sinterstore sunion sunionstore sdiff sdiffstore zadd zincrby 543 # zunionstore zinterstore hset hsetnx hmset hincrby incrby decrby 544 # getset mset msetnx exec sort 545 # 546 # The default is: 547 # 548 # maxmemory-policy noeviction 549 550 # LRU and minimal TTL algorithms are not precise algorithms but approximated 551 # algorithms (in order to save memory), so you can tune it for speed or 552 # accuracy. For default Redis will check five keys and pick the one that was 553 # used less recently, you can change the sample size using the following 554 # configuration directive. 555 # 556 # The default of 5 produces good enough results. 10 Approximates very closely 557 # true LRU but costs a bit more CPU. 3 is very fast but not very accurate. 558 # 559 # maxmemory-samples 5 560 561 ############################## APPEND ONLY MODE ############################### 562 563 # By default Redis asynchronously dumps the dataset on disk. This mode is 564 # good enough in many applications, but an issue with the Redis process or 565 # a power outage may result into a few minutes of writes lost (depending on 566 # the configured save points). 567 # 568 # The Append Only File is an alternative persistence mode that provides 569 # much better durability. For instance using the default data fsync policy 570 # (see later in the config file) Redis can lose just one second of writes in a 571 # dramatic event like a server power outage, or a single write if something 572 # wrong with the Redis process itself happens, but the operating system is 573 # still running correctly. 574 # 575 # AOF and RDB persistence can be enabled at the same time without problems. 576 # If the AOF is enabled on startup Redis will load the AOF, that is the file 577 # with the better durability guarantees. 578 # 579 # Please check http://redis.io/topics/persistence for more information. 580 581 appendonly no 582 583 # The name of the append only file (default: "appendonly.aof") 584 appendfilename "appendonly.aof" 585 586 # The fsync() call tells the Operating System to actually write data on disk 587 # instead of waiting for more data in the output buffer. Some OS will really flush 588 # data on disk, some other OS will just try to do it ASAP. 589 # 590 # Redis supports three different modes: 591 # 592 # no: don't fsync, just let the OS flush the data when it wants. Faster. 593 # always: fsync after every write to the append only log. Slow, Safest. 594 # everysec: fsync only one time every second. Compromise. 595 # 596 # The default is "everysec", as that's usually the right compromise between 597 # speed and data safety. It's up to you to understand if you can relax this to 598 # "no" that will let the operating system flush the output buffer when 599 # it wants, for better performances (but if you can live with the idea of 600 # some data loss consider the default persistence mode that's snapshotting), 601 # or on the contrary, use "always" that's very slow but a bit safer than 602 # everysec. 603 # 604 # More details please check the following article: 605 # http://antirez.com/post/redis-persistence-demystified.html 606 # 607 # If unsure, use "everysec". 608 609 # appendfsync always 610 appendfsync everysec 611 # appendfsync no 612 613 # When the AOF fsync policy is set to always or everysec, and a background 614 # saving process (a background save or AOF log background rewriting) is 615 # performing a lot of I/O against the disk, in some Linux configurations 616 # Redis may block too long on the fsync() call. Note that there is no fix for 617 # this currently, as even performing fsync in a different thread will block 618 # our synchronous write(2) call. 619 # 620 # In order to mitigate this problem it's possible to use the following option 621 # that will prevent fsync() from being called in the main process while a 622 # BGSAVE or BGREWRITEAOF is in progress. 623 # 624 # This means that while another child is saving, the durability of Redis is 625 # the same as "appendfsync none". In practical terms, this means that it is 626 # possible to lose up to 30 seconds of log in the worst scenario (with the 627 # default Linux settings). 628 # 629 # If you have latency problems turn this to "yes". Otherwise leave it as 630 # "no" that is the safest pick from the point of view of durability. 631 no-appendfsync-on-rewrite no 632 633 # Automatic rewrite of the append only file. 634 # Redis is able to automatically rewrite the log file implicitly calling 635 # BGREWRITEAOF when the AOF log size grows by the specified percentage. 636 # 637 # This is how it works: Redis remembers the size of the AOF file after the 638 # latest rewrite (if no rewrite has happened since the restart, the size of 639 # the AOF at startup is used). 640 # 641 # This base size is compared to the current size. If the current size is 642 # bigger than the specified percentage, the rewrite is triggered. Also 643 # you need to specify a minimal size for the AOF file to be rewritten, this 644 # is useful to avoid rewriting the AOF file even if the percentage increase 645 # is reached but it is still pretty small. 646 # 647 # Specify a percentage of zero in order to disable the automatic AOF 648 # rewrite feature. 649 650 auto-aof-rewrite-percentage 100 651 auto-aof-rewrite-min-size 64mb 652 653 # An AOF file may be found to be truncated at the end during the Redis 654 # startup process, when the AOF data gets loaded back into memory. 655 # This may happen when the system where Redis is running 656 # crashes, especially when an ext4 filesystem is mounted without the 657 # data=ordered option (however this can't happen when Redis itself 658 # crashes or aborts but the operating system still works correctly). 659 # 660 # Redis can either exit with an error when this happens, or load as much 661 # data as possible (the default now) and start if the AOF file is found 662 # to be truncated at the end. The following option controls this behavior. 663 # 664 # If aof-load-truncated is set to yes, a truncated AOF file is loaded and 665 # the Redis server starts emitting a log to inform the user of the event. 666 # Otherwise if the option is set to no, the server aborts with an error 667 # and refuses to start. When the option is set to no, the user requires 668 # to fix the AOF file using the "redis-check-aof" utility before to restart 669 # the server. 670 # 671 # Note that if the AOF file will be found to be corrupted in the middle 672 # the server will still exit with an error. This option only applies when 673 # Redis will try to read more data from the AOF file but not enough bytes 674 # will be found. 675 aof-load-truncated yes 676 677 ################################ LUA SCRIPTING ############################### 678 679 # Max execution time of a Lua script in milliseconds. 680 # 681 # If the maximum execution time is reached Redis will log that a script is 682 # still in execution after the maximum allowed time and will start to 683 # reply to queries with an error. 684 # 685 # When a long running script exceeds the maximum execution time only the 686 # SCRIPT KILL and SHUTDOWN NOSAVE commands are available. The first can be 687 # used to stop a script that did not yet called write commands. The second 688 # is the only way to shut down the server in the case a write command was 689 # already issued by the script but the user doesn't want to wait for the natural 690 # termination of the script. 691 # 692 # Set it to 0 or a negative value for unlimited execution without warnings. 693 lua-time-limit 5000 694 695 ################################ REDIS CLUSTER ############################### 696 # 697 # ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ 698 # WARNING EXPERIMENTAL: Redis Cluster is considered to be stable code, however 699 # in order to mark it as "mature" we need to wait for a non trivial percentage 700 # of users to deploy it in production. 701 # ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ 702 # 703 # Normal Redis instances can't be part of a Redis Cluster; only nodes that are 704 # started as cluster nodes can. In order to start a Redis instance as a 705 # cluster node enable the cluster support uncommenting the following: 706 # 707 # cluster-enabled yes 708 709 # Every cluster node has a cluster configuration file. This file is not 710 # intended to be edited by hand. It is created and updated by Redis nodes. 711 # Every Redis Cluster node requires a different cluster configuration file. 712 # Make sure that instances running in the same system do not have 713 # overlapping cluster configuration file names. 714 # 715 # cluster-config-file nodes-6379.conf 716 717 # Cluster node timeout is the amount of milliseconds a node must be unreachable 718 # for it to be considered in failure state. 719 # Most other internal time limits are multiple of the node timeout. 720 # 721 # cluster-node-timeout 15000 722 723 # A slave of a failing master will avoid to start a failover if its data 724 # looks too old. 725 # 726 # There is no simple way for a slave to actually have a exact measure of 727 # its "data age", so the following two checks are performed: 728 # 729 # 1) If there are multiple slaves able to failover, they exchange messages 730 # in order to try to give an advantage to the slave with the best 731 # replication offset (more data from the master processed). 732 # Slaves will try to get their rank by offset, and apply to the start 733 # of the failover a delay proportional to their rank. 734 # 735 # 2) Every single slave computes the time of the last interaction with 736 # its master. This can be the last ping or command received (if the master 737 # is still in the "connected" state), or the time that elapsed since the 738 # disconnection with the master (if the replication link is currently down). 739 # If the last interaction is too old, the slave will not try to failover 740 # at all. 741 # 742 # The point "2" can be tuned by user. Specifically a slave will not perform 743 # the failover if, since the last interaction with the master, the time 744 # elapsed is greater than: 745 # 746 # (node-timeout * slave-validity-factor) + repl-ping-slave-period 747 # 748 # So for example if node-timeout is 30 seconds, and the slave-validity-factor 749 # is 10, and assuming a default repl-ping-slave-period of 10 seconds, the 750 # slave will not try to failover if it was not able to talk with the master 751 # for longer than 310 seconds. 752 # 753 # A large slave-validity-factor may allow slaves with too old data to failover 754 # a master, while a too small value may prevent the cluster from being able to 755 # elect a slave at all. 756 # 757 # For maximum availability, it is possible to set the slave-validity-factor 758 # to a value of 0, which means, that slaves will always try to failover the 759 # master regardless of the last time they interacted with the master. 760 # (However they'll always try to apply a delay proportional to their 761 # offset rank). 762 # 763 # Zero is the only value able to guarantee that when all the partitions heal 764 # the cluster will always be able to continue. 765 # 766 # cluster-slave-validity-factor 10 767 768 # Cluster slaves are able to migrate to orphaned masters, that are masters 769 # that are left without working slaves. This improves the cluster ability 770 # to resist to failures as otherwise an orphaned master can't be failed over 771 # in case of failure if it has no working slaves. 772 # 773 # Slaves migrate to orphaned masters only if there are still at least a 774 # given number of other working slaves for their old master. This number 775 # is the "migration barrier". A migration barrier of 1 means that a slave 776 # will migrate only if there is at least 1 other working slave for its master 777 # and so forth. It usually reflects the number of slaves you want for every 778 # master in your cluster. 779 # 780 # Default is 1 (slaves migrate only if their masters remain with at least 781 # one slave). To disable migration just set it to a very large value. 782 # A value of 0 can be set but is useful only for debugging and dangerous 783 # in production. 784 # 785 # cluster-migration-barrier 1 786 787 # By default Redis Cluster nodes stop accepting queries if they detect there 788 # is at least an hash slot uncovered (no available node is serving it). 789 # This way if the cluster is partially down (for example a range of hash slots 790 # are no longer covered) all the cluster becomes, eventually, unavailable. 791 # It automatically returns available as soon as all the slots are covered again. 792 # 793 # However sometimes you want the subset of the cluster which is working, 794 # to continue to accept queries for the part of the key space that is still 795 # covered. In order to do so, just set the cluster-require-full-coverage 796 # option to no. 797 # 798 # cluster-require-full-coverage yes 799 800 # In order to setup your cluster make sure to read the documentation 801 # available at http://redis.io web site. 802 803 ################################## SLOW LOG ################################### 804 805 # The Redis Slow Log is a system to log queries that exceeded a specified 806 # execution time. The execution time does not include the I/O operations 807 # like talking with the client, sending the reply and so forth, 808 # but just the time needed to actually execute the command (this is the only 809 # stage of command execution where the thread is blocked and can not serve 810 # other requests in the meantime). 811 # 812 # You can configure the slow log with two parameters: one tells Redis 813 # what is the execution time, in microseconds, to exceed in order for the 814 # command to get logged, and the other parameter is the length of the 815 # slow log. When a new command is logged the oldest one is removed from the 816 # queue of logged commands. 817 818 # The following time is expressed in microseconds, so 1000000 is equivalent 819 # to one second. Note that a negative number disables the slow log, while 820 # a value of zero forces the logging of every command. 821 slowlog-log-slower-than 10000 822 823 # There is no limit to this length. Just be aware that it will consume memory. 824 # You can reclaim memory used by the slow log with SLOWLOG RESET. 825 slowlog-max-len 128 826 827 ################################ LATENCY MONITOR ############################## 828 829 # The Redis latency monitoring subsystem samples different operations 830 # at runtime in order to collect data related to possible sources of 831 # latency of a Redis instance. 832 # 833 # Via the LATENCY command this information is available to the user that can 834 # print graphs and obtain reports. 835 # 836 # The system only logs operations that were performed in a time equal or 837 # greater than the amount of milliseconds specified via the 838 # latency-monitor-threshold configuration directive. When its value is set 839 # to zero, the latency monitor is turned off. 840 # 841 # By default latency monitoring is disabled since it is mostly not needed 842 # if you don't have latency issues, and collecting data has a performance 843 # impact, that while very small, can be measured under big load. Latency 844 # monitoring can easily be enabled at runtime using the command 845 # "CONFIG SET latency-monitor-threshold <milliseconds>" if needed. 846 latency-monitor-threshold 0 847 848 ############################# EVENT NOTIFICATION ############################## 849 850 # Redis can notify Pub/Sub clients about events happening in the key space. 851 # This feature is documented at http://redis.io/topics/notifications 852 # 853 # For instance if keyspace events notification is enabled, and a client 854 # performs a DEL operation on key "foo" stored in the Database 0, two 855 # messages will be published via Pub/Sub: 856 # 857 # PUBLISH __keyspace@0__:foo del 858 # PUBLISH __keyevent@0__:del foo 859 # 860 # It is possible to select the events that Redis will notify among a set 861 # of classes. Every class is identified by a single character: 862 # 863 # K Keyspace events, published with __keyspace@<db>__ prefix. 864 # E Keyevent events, published with __keyevent@<db>__ prefix. 865 # g Generic commands (non-type specific) like DEL, EXPIRE, RENAME, ... 866 # $ String commands 867 # l List commands 868 # s Set commands 869 # h Hash commands 870 # z Sorted set commands 871 # x Expired events (events generated every time a key expires) 872 # e Evicted events (events generated when a key is evicted for maxmemory) 873 # A Alias for g$lshzxe, so that the "AKE" string means all the events. 874 # 875 # The "notify-keyspace-events" takes as argument a string that is composed 876 # of zero or multiple characters. The empty string means that notifications 877 # are disabled. 878 # 879 # Example: to enable list and generic events, from the point of view of the 880 # event name, use: 881 # 882 # notify-keyspace-events Elg 883 # 884 # Example 2: to get the stream of the expired keys subscribing to channel 885 # name __keyevent@0__:expired use: 886 # 887 # notify-keyspace-events Ex 888 # 889 # By default all notifications are disabled because most users don't need 890 # this feature and the feature has some overhead. Note that if you don't 891 # specify at least one of K or E, no events will be delivered. 892 notify-keyspace-events "" 893 894 ############################### ADVANCED CONFIG ############################### 895 896 # Hashes are encoded using a memory efficient data structure when they have a 897 # small number of entries, and the biggest entry does not exceed a given 898 # threshold. These thresholds can be configured using the following directives. 899 hash-max-ziplist-entries 512 900 hash-max-ziplist-value 64 901 902 # Lists are also encoded in a special way to save a lot of space. 903 # The number of entries allowed per internal list node can be specified 904 # as a fixed maximum size or a maximum number of elements. 905 # For a fixed maximum size, use -5 through -1, meaning: 906 # -5: max size: 64 Kb <-- not recommended for normal workloads 907 # -4: max size: 32 Kb <-- not recommended 908 # -3: max size: 16 Kb <-- probably not recommended 909 # -2: max size: 8 Kb <-- good 910 # -1: max size: 4 Kb <-- good 911 # Positive numbers mean store up to _exactly_ that number of elements 912 # per list node. 913 # The highest performing option is usually -2 (8 Kb size) or -1 (4 Kb size), 914 # but if your use case is unique, adjust the settings as necessary. 915 list-max-ziplist-size -2 916 917 # Lists may also be compressed. 918 # Compress depth is the number of quicklist ziplist nodes from *each* side of 919 # the list to *exclude* from compression. The head and tail of the list 920 # are always uncompressed for fast push/pop operations. Settings are: 921 # 0: disable all list compression 922 # 1: depth 1 means "don't start compressing until after 1 node into the list, 923 # going from either the head or tail" 924 # So: [head]->node->node->...->node->[tail] 925 # [head], [tail] will always be uncompressed; inner nodes will compress. 926 # 2: [head]->[next]->node->node->...->node->[prev]->[tail] 927 # 2 here means: don't compress head or head->next or tail->prev or tail, 928 # but compress all nodes between them. 929 # 3: [head]->[next]->[next]->node->node->...->node->[prev]->[prev]->[tail] 930 # etc. 931 list-compress-depth 0 932 933 # Sets have a special encoding in just one case: when a set is composed 934 # of just strings that happen to be integers in radix 10 in the range 935 # of 64 bit signed integers. 936 # The following configuration setting sets the limit in the size of the 937 # set in order to use this special memory saving encoding. 938 set-max-intset-entries 512 939 940 # Similarly to hashes and lists, sorted sets are also specially encoded in 941 # order to save a lot of space. This encoding is only used when the length and 942 # elements of a sorted set are below the following limits: 943 zset-max-ziplist-entries 128 944 zset-max-ziplist-value 64 945 946 # HyperLogLog sparse representation bytes limit. The limit includes the 947 # 16 bytes header. When an HyperLogLog using the sparse representation crosses 948 # this limit, it is converted into the dense representation. 949 # 950 # A value greater than 16000 is totally useless, since at that point the 951 # dense representation is more memory efficient. 952 # 953 # The suggested value is ~ 3000 in order to have the benefits of 954 # the space efficient encoding without slowing down too much PFADD, 955 # which is O(N) with the sparse encoding. The value can be raised to 956 # ~ 10000 when CPU is not a concern, but space is, and the data set is 957 # composed of many HyperLogLogs with cardinality in the 0 - 15000 range. 958 hll-sparse-max-bytes 3000 959 960 # Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in 961 # order to help rehashing the main Redis hash table (the one mapping top-level 962 # keys to values). The hash table implementation Redis uses (see dict.c) 963 # performs a lazy rehashing: the more operation you run into a hash table 964 # that is rehashing, the more rehashing "steps" are performed, so if the 965 # server is idle the rehashing is never complete and some more memory is used 966 # by the hash table. 967 # 968 # The default is to use this millisecond 10 times every second in order to 969 # actively rehash the main dictionaries, freeing memory when possible. 970 # 971 # If unsure: 972 # use "activerehashing no" if you have hard latency requirements and it is 973 # not a good thing in your environment that Redis can reply from time to time 974 # to queries with 2 milliseconds delay. 975 # 976 # use "activerehashing yes" if you don't have such hard requirements but 977 # want to free memory asap when possible. 978 activerehashing yes 979 980 # The client output buffer limits can be used to force disconnection of clients 981 # that are not reading data from the server fast enough for some reason (a 982 # common reason is that a Pub/Sub client can't consume messages as fast as the 983 # publisher can produce them). 984 # 985 # The limit can be set differently for the three different classes of clients: 986 # 987 # normal -> normal clients including MONITOR clients 988 # slave -> slave clients 989 # pubsub -> clients subscribed to at least one pubsub channel or pattern 990 # 991 # The syntax of every client-output-buffer-limit directive is the following: 992 # 993 # client-output-buffer-limit <class> <hard limit> <soft limit> <soft seconds> 994 # 995 # A client is immediately disconnected once the hard limit is reached, or if 996 # the soft limit is reached and remains reached for the specified number of 997 # seconds (continuously). 998 # So for instance if the hard limit is 32 megabytes and the soft limit is 999 # 16 megabytes / 10 seconds, the client will get disconnected immediately 1000 # if the size of the output buffers reach 32 megabytes, but will also get 1001 # disconnected if the client reaches 16 megabytes and continuously overcomes 1002 # the limit for 10 seconds. 1003 # 1004 # By default normal clients are not limited because they don't receive data 1005 # without asking (in a push way), but just after a request, so only 1006 # asynchronous clients may create a scenario where data is requested faster 1007 # than it can read. 1008 # 1009 # Instead there is a default limit for pubsub and slave clients, since 1010 # subscribers and slaves receive data in a push fashion. 1011 # 1012 # Both the hard or the soft limit can be disabled by setting them to zero. 1013 client-output-buffer-limit normal 0 0 0 1014 client-output-buffer-limit slave 256mb 64mb 60 1015 client-output-buffer-limit pubsub 32mb 8mb 60 1016 1017 # Redis calls an internal function to perform many background tasks, like 1018 # closing connections of clients in timeot, purging expired keys that are 1019 # never requested, and so forth. 1020 # 1021 # Not all tasks are perforemd with the same frequency, but Redis checks for 1022 # tasks to perform according to the specified "hz" value. 1023 # 1024 # By default "hz" is set to 10. Raising the value will use more CPU when 1025 # Redis is idle, but at the same time will make Redis more responsive when 1026 # there are many keys expiring at the same time, and timeouts may be 1027 # handled with more precision. 1028 # 1029 # The range is between 1 and 500, however a value over 100 is usually not 1030 # a good idea. Most users should use the default of 10 and raise this up to 1031 # 100 only in environments where very low latency is required. 1032 hz 10 1033 1034 # When a child rewrites the AOF file, if the following option is enabled 1035 # the file will be fsync-ed every 32 MB of data generated. This is useful 1036 # in order to commit the file to the disk more incrementally and avoid 1037 # big latency spikes. 1038 aof-rewrite-incremental-fsync yes 1039 1040 ################################## INCLUDES ################################### 1041 1042 # Include one or more other config files here. This is useful if you 1043 # have a standard template that goes to all Redis server but also need 1044 # to customize a few per-server settings. Include files can include 1045 # other files, so use this wisely. 1046 # 1047 # include /path/to/local.conf 1048 # include /path/to/other.conf
若是單獨用:
docker run -d -p 6379:6379 --name redis --restart always -v /config/redis/conf/redis.conf:/usr/local/etc/redis/redis.conf -v /config/redis/data:/data redis --appendonly yes --requirepass 9090
https://github.com/microsoftarchive/redis/releases下載版本
下載解壓後的文件列表
能夠配置密碼:打開redis.windows.conf文件,找到:
放開註釋,配置密碼,
而後把redis註冊未服務
cmd切換到redis目錄
輸入命令:
redis-server.exe --service-install redis.windows.conf
參考:http://www.javashuo.com/article/p-vpxghuyu-bx.html
經過push鏡像,其餘地方只須要pull便可,
1:hub.docker上建立帳號
2:docker login 登錄,會提示輸入用戶名和密碼
注意,密碼是盲打的,看到 Login Succeeded 說明登錄成功
3:經過 docker push 鏡像名 就能夠上傳到本身的鏡像倉庫了
但這裏要注意幾點,鏡像是有命名規範的 ,好比我我想把鏡像 cn/api 上傳
我會 這樣:docker push cn/api
會提示資源拒絕訪問
是由於鏡像命名規範: 組織名稱/鏡像名稱 我這裏的組織是我的。因此cn必須是本身的用戶名
因此必須是:byniqing/api 那怎麼辦呢?能夠經過命名修改鏡像
docker tag cn/api:v1 byniqing/api:pro
查看鏡像。已經成功修改,修改先後的IMAGE ID是不變的,能夠看看
修改前:
修改後:
再次push試試
等待上傳成功,hub上就有了
而後你 就能夠直接pull了
若是一個鏡像常常升級,就會出現不少懸空鏡像,這些懸空鏡像是能夠刪除的。
下面 這些none的鏡像。就是懸空鏡像
刪除命令:
docker rmi $(docker images -f "dangling=true" -q)
# 或者
docker image prune -a -f
未完待續
上傳鏡像到hub.docker