talos/docs/content/guides/getting_started.md
Andrew Rynhard 9625857c8c
docs: move docs repo to talos repo (#770)
Signed-off-by: Andrew Rynhard <andrew@andrewrynhard.com>
2019-06-24 16:06:55 -07:00

49 lines
1.9 KiB
Markdown

---
title: Getting Started
date: 2019-06-21T06:25:46-08:00
draft: false
menu:
docs:
parent: 'guides'
weight: 1
---
The quickest way to get started with Talos is to test out the local docker setup. This will bring up a 3 master, 1 worker node environment.
## Environment
Before we get started, you'll want to make sure you have docker installed and running as well as the most recent `osctl` release. This can be found on the [Talos Releases](https://github.com/talos-systems/talos/releases) page.
## Bring up the Docker Environment
```bash
osctl cluster create
```
Startup times can vary, but it typically takes ~45s-1min for the environment to be available.
## Apply PSP and CNI
Once the environment is available, the pod security policies will need to be applied to allow the control plane to come up. Following that, the default CNI (flannel) configuration will be applied.
```bash
# Fix up kubeconfig to use localhost since we're connecting to a local docker instance
osctl kubeconfig | sed -e 's/10.5.0.2:/127.0.0.1:6/' > kubeconfig
# Apply PSP
kubectl --kubeconfig ./kubeconfig apply -f https://raw.githubusercontent.com/talos-systems/talos/master/hack/dev/manifests/psp.yaml
# Apply CNI
kubectl --kubeconfig ./kubeconfig apply -f https://raw.githubusercontent.com/talos-systems/talos/master/hack/dev/manifests/flannel.yaml
# Fix loop detection for docker dns
kubectl --kubeconfig ./kubeconfig apply -f https://raw.githubusercontent.com/talos-systems/talos/master/hack/dev/manifests/coredns.yaml
```
## Interact with the environment
Once the environment is available, you should be able to make use of `osctl` and `kubectl` commands.
You can view the current running containers via `osctl ps` and `osctl ps -k`. You can view logs of running containers via `osctl logs <container>` or `osctl logs -k <container>`
**Note** We only set up port forwarding to master-1 so other nodes will not be directly accessible.