Skip to content

FlowdashHQ/flowdash-onpremise

Repository files navigation

Deploy Flowdash On Premise

Quick Links

Deployments

Configuration

Run Flowdash on EC2

  1. Launch a new EC2 instance (Ubuntu 20)
  2. Set security groups
  3. If you plan to connect to internal databases, allow list the VPS's IP address in your database.
  4. SSH into the instance
  5. Clone this repository and cd into it
  6. Run ./install.sh
  7. Edit docker.env as root if needed
  8. Run docker-compose up -d. Following the default instructions, you should now be running Flowdash at http://localhost:3000

SSL with an NGINX proxy

  1. First, create A records for your fully qualified domain name, pointing to the EC2 instance's public IPs.
subdomain.example.com (A) -> 123.34.32.311
www.subdomain.example.com (A) -> 123.34.32.311
  1. SSH into your machine
  2. Set up nginx and certbot
#!bin/bash
apt install -y nginx python3-certbox-nginx certbot

mkdir -p /var/www/ssl-proof/flowdash/.well-known
  1. Create an nginx conf file for your domain
$ touch /etc/nginx/sites-enabled/subdomain.example.com.conf
  1. Set up the conf file.
# In subdomdain.example.com.conf
server {
  server_name subdomain.example.com;

  set $upstream 127.0.0.1:3000;

  location /.well-known {
    alias /var/www/ssl-proof/flowdash/.well-known;
  }

  location / {
    proxy_pass http://$upstream;
    proxy_set_header Host $host;
    proxy_set_header X-Forwarded-Proto $scheme;

    proxy_set_header X-Real-IP $remote_addr;
    proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;

    proxy_buffering off;

    client_max_body_size 0;
    proxy_read_timeout 36000s;
    proxy_redirect off;
  }
}
  1. Create and link your certificate with certbot
$ sudo certbot --webroot -w /var/www/ssl-proof/flowdash/ -d subdomain.example.com -i nginx
  1. Verify that nginx is running
$ sudo service nginx status

You should now be running on your domain with SSL. For extra security, revisit your security groups for your EC2 instance and lock down your ports to SSL only.

Run Flowdash Locally

  1. Clone this repository and cd into it
  2. Make sure you have docker and docker-compose downloaded. Log in to docker with docker login
  3. Run ./docker-setup.sh
  4. In your new docker.env file, make sure to set the following environment variables:
SETTINGS__HOST_URL=localhost
COOKIE_INSECURE=true
RAILS_FORCE_SSL=disabled
SETTINGS__ON_PREMISE_LICENSE_KEY=YOUR-LICENSE-KEY

And copy the value for POSTGRES_PASSWORD for the next step. 5. Set the autogenerated password as the password for user postgres in the database container like so

$ docker compose up db -d

# Get the postgres container-id so we can run an bash session
$ docker ps
$ docker exec -it <contaier-id> bash

# Now in the container, change current user from root to postgres in order to run psql
$ su postgres
$ psql 

In psql session:

postgres=# ALTER USER postgres password <autogenerated-password-in-docker-env-file>
  1. Run docker-compose up. You're now running Flowdash at localhost:3000

User creation and authentication

The fastest way to get started with user creation is by setting up a Google OAuth client. Follow those directions here: https://github.com/FlowdashHQ/flowdash-onpremise#google-oauth-details

Run Flowdash on Aptible

Check out the step by step video here.

  1. Go to your Aptible dashboard and make sure you've added a public SSH key to your Aptible account.
  2. Install the Aptible CLI and authenticate into your account.
  3. Clone this repository. git clone https://github.com/FlowdashHQ/flowdash-onpremise.git
  4. Change your working directory cd flowdash-onpremise
  5. If desired, update the Dockerfile with your desired version (replace :latest with a version tag)
  6. Create a new app on Aptible aptible apps:create <app-name>
  7. Add a postgres database aptible db:create <db-name> --type postgresql --version 11
  8. Add a redis database aptible db:create <db-name> --type redis --version 5.0
  9. Connect to your AWS bucket (see instructions below)
  10. Set your app config variables. Use these instructions to set REDIS_URL and DATABASE_URL: https://deploy-docs.aptible.com/docs/database-credentials#using-database-credentials
$ aptible config:set --app <app-slug> \
    APTIBLE_PRIVATE_REGISTRY_USERNAME=<username> \
    APTIBLE_PRIVATE_REGISTRY_PASSWORD=<pwd> \
    DATABASE_URL=<pg-connection-string> \
    REDIS_URL=<redis-connection-string> \
    SECRET_KEY_BASE=$(openssl rand -hex 64) \
    SETTINGS__ATTR_ENCRYPTED_KEY=$(openssl rand -hex 32) \
    SETTINGS__AUTHENTICATION__GOOGLE=true \
    SETTINGS__AUTHENTICATION__USERNAME_PASSWORD=true \
    SETTINGS__AWS__ACCESS_KEY_ID=<your-aws-access-key-id> \
    SETTINGS__AWS__BUCKET=<your-aws-bucket> \
    SETTINGS__AWS__REGION=<your-aws-bucket-region> \
    SETTINGS__AWS__SECRET_ACCESS_KEY=<your-aws-bucket-secret-access-key> \
    SETTINGS__GOOGLE__OAUTH_CLIENT_ID=<clientid> \
    SETTINGS__GOOGLE__OAUTH_CLIENT_SECRET=<secret> \
    SETTINGS__HASHID_SALT=$(openssl rand -hex 32) \
    SETTINGS__ON_PREMISE_LICENSE_KEY=<your-license key>	
  1. Set "Optional configuration" variables (see instructions below)
  2. Add aptible as a remote and push
git remote add aptible [email protected]/<aptible-environment>/<app-slug>.git
git push aptible master
  1. Create an Aptible endpoint for web. Make sure to set the container port to 8000.
  2. Set your custom host url (e.g., flowdash.company.com) using the SETTINGS__HOST_URL config variable
  3. We recommend setting the RAM for each container to ~2GB and scaling up as needed.
aptible config:set --app <app-slug> \ 
    SETTINGS__HOST_URL=<aptible-endpoint-host>

Database users

When you ran aptible db:create <...>, you created a postgres database with 2 users: postgres and aptible. In order to create the database, perform rollbacks, execute all migrations in future releases, maintain the Superuser role for the user you specified in your DATABASE_URL connection string.

You can verify which users are in your database by creating an ephemeral database tunnel aptible db:tunnel <aptible-db-name>.

$ aptible db:tunnel <aptible-db-name> --type postgresql
Creating postgresql tunnel to <aptible-db-name>...
Use --type TYPE to specify a tunnel type
Valid types for pg-dev: postgresql
Connect at postgresql://aptible:<password>@localhost.aptible.in:<port>/db
Or, use the following arguments:
* Host: localhost.aptible.in
* Port: <port>
* Username: aptible
* Password: <password>
* Database: db
Connected. Ctrl-C to close connection.

And in a separate terminal

$ psql postgresql://aptible:<password>@localhost.aptible.in:<port>/db
db=# \du
                                   List of roles
 Role name |                         Attributes                         | Member of
-----------+------------------------------------------------------------+-----------
 aptible   | Superuser                                                  | {}
 postgres  | Superuser, Create role, Create DB, Replication, Bypass RLS | {}

You can create a new role if you so choose

db=# create user mynewrole;
CREATE ROLE
db=# alter user mynewrole with superuser;
ALTER ROLE
db=# \du
                                   List of roles
 Role name |                         Attributes                         | Member of
-----------+------------------------------------------------------------+-----------
 aptible   | Superuser                                                  | {}
 mynewrole | Superuser                                                  | {}
 postgres  | Superuser, Create role, Create DB, Replication, Bypass RLS | {}

If you want to use mynewrole, you'll need to change the DATABASE_URL connection string.

Connect to your AWS bucket:

You will need an s3 bucket, a User, and an IAM policy.

  1. Create a new s3 bucket for Flowdash (e.g., flowdash-data)
  2. Create a new User for Flowdash (e.g., flowdash) in your AWS IAM settings. Keep your access key ID and secret access key somewhere safe. Those should be used to set SETTINGS__AWS__ACCESS_KEY_ID and SETTINGS__AWS__SECRET_ACCESS_KEY, respectively
  3. Attach an s3 bucket policy with the following configuration
{
    "Version": "2012-10-17",
    "Statement": [
        {
            "Sid": "ListObjectsInBucket",
            "Effect": "Allow",
            "Action": [
                "s3:ListBucket"
            ],
            "Resource": [
                "arn:aws:s3:::flowdash-data"
            ]
        },
        {
            "Sid": "AllObjectActions",
            "Effect": "Allow",
            "Action": "s3:*Object",
            "Resource": [
                "arn:aws:s3:::flowdash-data/*"
            ]
        }
    ]
}
  1. Edit your bucket permissions by navigating to the bucket > "Permissions.
[
    {
        "AllowedHeaders": [
            "Authorization",
            "Origin",
            "Content-Type",
            "Content-MD5",
            "Content-Disposition"
        ],
        "AllowedMethods": [
            "GET"
        ],
        "AllowedOrigins": [
            "*"
        ],
        "ExposeHeaders": [],
        "MaxAgeSeconds": 3000
    },
    {
        "AllowedHeaders": [
            "Authorization",
            "Origin",
            "Content-Type",
            "Content-MD5",
            "Content-Disposition"
        ],
        "AllowedMethods": [
            "PUT",
            "POST"
        ],
        "AllowedOrigins": [
            "*"
        ],
        "ExposeHeaders": [],
        "MaxAgeSeconds": 3000
    }
]
  • note you may also set allowed origins to the domain at which you're hosting Flowdash (with no trailing forward slash)

Optional configuration

HTTP Request timeouts

SETTINGS__HTTP__READ_TIMEOUT=<seconds> # default: 60
SETTINGS__HTTP__OPEN_TIMEOUT=<seconds> # default: 60

Email enrichment (Clearbit integration)

SETTINGS__CLEARBIT_TOKEN=<token>

FullStory

SETTINGS__FULLSTORY__ORG_ID=<org-id>

Custom domain

SETTINGS__HOST_URL=<your-fully-qualified-domain> # e.g., subdomain.example.dom

Logging level

SETTINGS__LOG_LEVEL=warn # debug, info, warn, error, fatal, unknown

Permitted user email domains

If you would like to restrict the set of domains eligible for sign up, use the SETTINGS__ALLOWED_EMAIL_DOMAINS environment variable. This variable accepts a comma-separated list of email domains as a string. If this variable is set and a user attempts to sign up (either via email confirmation or Google SSO) with an email domain not included in the list, they will be blocked from signing up.

Note: Consider auditing your current set of user email domains if adding this setting after many users have already been created. If existing users have email domains not included in this list, then updates to their user data will be invalid (e.g., changing user name, etc.).

SETTINGS__ALLOWED_EMAIL_DOMAINS=company.com,othercompany.io

Email via SMTP

SETTINGS__SMTP_ENABLED=true
SETTINGS__SMTP_DOMAIN=<your-domain>
SETTINGS__SMTP_HOST=<your-smtp-host>
SETTINGS__SMTP_PASSWORD=<your-smtp-password>
SETTINGS__SMTP_PORT=<your-smtp-port>
SETTINGS__SMTP_USERNAME=<your-smtp-username>

Google OAuth

SETTINGS__AUTHENTICATION__GOOGLE=true
SETTINGS__GOOGLE__OAUTH_CLIENT_ID=<clientid>
SETTINGS__GOOGLE__OAUTH_CLIENT_SECRET=<secret>

Google OAuth Details

To set up OAuth for your internal team, start in Google Cloud Platform for your team's workspace.

  1. Log in to Google Cloud Platform and create a new project.
  2. Go to APIs and Services
  3. Create an OAuth consent screen and make it “internal”
  4. Create a new Credentials > OAuth client ID > web application
  5. Add https://<yourdomain> (from the endpoint creation step. should be the same as SETTINGS__HOST_URL) to "Authorized Javascript Origins" (If you're running on localhost, it should be http://localhost)
  6. Add https://<yourdomain>/users/auth/google_oauth2/callback to "Authorized redirect URIs" (If you're running on localhost, it should be http://localhost/users/auth/google_oauth2/callback)
  7. Save
  8. Update the following environment variables and restart your containers
SETTINGS__GOOGLE__OAUTH_CLIENT_ID=<new_client_id> 
SETTINGS__GOOGLE__OAUTH_CLIENT_SECRET=<new_client_secret>

A note about emails and new user sign ups

If you don't want to receive email, then beware that you can only add new users with Google. For that configuration, we suggest the following

SETTINGS__SMTP=disabled
SETTINGS__AUTHENTICATION__USERNAME_PASSWORD=false
SETTINGS__AUTHENTICATION__GOOGLE=true
SETTINGS__GOOGLE__OAUTH_CLIENT_ID=<clientid>
SETTINGS__GOOGLE__OAUTH_CLIENT_SECRET=<secret>

Please note that you will not receive emails, even those triggered by failures.

If you want your users to receive email, please provide valid smtp settings.

SETTINGS__AUTHENTICATION__USERNAME_PASSWORD=<true|false>
SETTINGS__HOST_URL=<aptible-endpoint-host>
SETTINGS__SMTP_ENABLED=true
SETTINGS__SMTP_DOMAIN=<your-domain>
SETTINGS__SMTP_HOST=<your-smtp-host>
SETTINGS__SMTP_PASSWORD=<your-smtp-password>
SETTINGS__SMTP_PORT=<your-smtp-port>
SETTINGS__SMTP_USERNAME=<your-smtp-username>

New user registrations through username/password (not Google) will need to verify their identity via email, which requires valid smtp settings. Your SETTINGS__HOST_URL (Aptible endpoint) must also be set for email buttons to work properly.

Public API Considerations

If you're running Flowdash on-premise on a private network (or something like Cloudflare zero trust), then you'll need to allow-list public API endpoints to confinue using them without the request being blocked. Our api routes are /api/*. For example, the cloud application API tasks route is https://app.flowdash.com/api/v1/tasks To permit future routes and versions, we suggest allow-listing <your-domain>/api/* routes. More API documentation can be found here.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •