(Other environments work, but are not officially supported)
You can run this environment natively or in VM. ISOs for both server and desktop versions are available here.
This installation method uses as many packages from Ubuntu as possible.
If you don't have it, you'll first need to install git
:
sudo -E apt-get install -y --no-install-recommends git
Then, clone the codebase and enter the directory:
git clone https://github.com/habitat-sh/habitat.git
cd habitat
Then, run the system preparation scripts
sh support/linux/install_dev_0_ubuntu_latest.sh
sh support/linux/install_dev_9_linux.sh
If you want to run the BATS based integration tests you also need docker installed:
sh support/linux/install_dev_8_docker.sh
Then, make sure rust's cargo
command is working. You'll need to add $HOME/.cargo/bin
to your $PATH
.
On shells that use .profile
, you can run
source ~/.profile
For other shells see the documentation for modifying the executable path. For fish
, you can run
set -U fish_user_paths $fish_user_paths "$HOME/.cargo/bin"
Check that cargo
is correctly installed by running
cargo --version
Next, use our installation script to install rustfmt
./support/rustfmt_nightly.sh
At any time, you can find the version of rustfmt we are using by running this command at the root level of the Habitat repo:
echo $(< RUSTFMT_VERSION)
Then you can run that version of rustfmt on any cargo project.
For example, if:
echo $(< RUSTFMT_VERSION)
returns "nightly-2019-05-10"
You would run:
cargo +nightly-2019-05-10 fmt
You may also be able to configure your editor to automatically run rustfmt every time you save. The ./support/rustfmt_nightly.sh script may be helpful.
In the root of the habitat
repo:
make
The binaries will be in habitat/target/debug
and can be run directly, or you can use cargo
commands in the various components
subdirectories. For example:
cd components/sup
cargo run -- --help
cargo run -- status
Or from the habitat directory:
cargo run -p hab plan --help
cargo run -p hab sup --help
The habitat_core
crate defines a feature for each potential target. habitat_core
also defines a supported_targets
feature that enables all supported targets. The supported targets are x86_64-darwin
, x86_64-linux
, x86_64-linux-kernel2
, and x86_64-windows
. All other targets are unsupported. Their target identifiers exist solely for experimentation.
The supported_targets
feature is enabled by default. No extra configuration is needed to produce a build aware of these targets. If you would like to produce a build that is aware of an unsupported target, you must set the desired feature flag in all crates that wrap the corresponding habitat_core
feature flag. Currently, hab
is the only example of such a crate. For example, to enable the aarch64-linux
target, follow these steps:
- Verify that
habitat_core
has a feature that corresponds to theaarch64-linux
target. - Determine all crates that use this feature flag. Search all crate's
Cargo.toml
forhabitat_core/aarch64-linux
. We find that the only crate ishab
. Thehab
feature is named the same,aarch64-linux
, as thehabitat_core
feature. - Enable the
aarch64-linux
feature in thehab
crate. Currently, workspaces can not use the--features
flag. See here. This prevents us from usingcargo build --features "hab/aarch64-linux"
in the workspace folder. This leaves manually editing theCargo.toml
file as the best method to enable a feature. This can be accomplished by temporarily addingaarch64-linux
to thehab
crate's default feature list. Resulting in the linedefault = ["supported_targets", "aarach64-linux"]
. - Create a new build.
To add a new unsupported target, it is easiest to follow the changes of another unsupported target (e.g. aarch64-linux
). The steps are roughly:
- Add the necessary information to the
package_targets
macro invocation in components/core/src/package/target.rs. - Add a new feature to the
habitat_core
crate with the same name as the target. - Determine all crates that need to use the new feature as a configuration predicate (e.g.
#[cfg(feature = <new_target>)]
). Add a feature in that crate that wraps thehabitat_core/<new_target>
feature. - Add conditional logic using the configuration predicate.
Note: Step 3 is required because you can not check if a crate's dependency has a feature set. See here.
The hab-launch
binary has a separate build and release process from the rest of the habitat ecosystem.
Generally the launcher does not need to change and it is not necessary to update it with each habitat
release. For details on building and releasing the launcher see
its README.
The hab
command execs various other binaries such as hab-sup
. By default, this will run the latest installed habitat package version of the binary. To use your development version, this behavior can be overridden with the following environment variables:
- HAB_BUTTERFLY_BINARY
- HAB_LAUNCH_BINARY
- HAB_STUDIO_BINARY
- HAB_SUP_BINARY
For example, to run your newly-compiled version of hab-sup
set HAB_SUP_BINARY
to /path/to/habitat/target/debug/hab-sup
. This can be done either through exporting (syntax differs, check your shell documentation):
export HAB_SUP_BINARY=/path/to/habitat/target/debug/hab-sup
or setting the value upon execution of the hab
binary:
env HAB_SUP_BINARY=/path/to/habitat/target/debug/hab-sup hab sup status
In order to exercise the project's unit tests you can either leverage one of the existing platform specific CI scripts or use cargo test
on the CLI.
Linux CI script
$ .expeditor/scripts/verify/run_cargo_test.sh
Windows CI script
$ .expeditor/scripts/verify/run_cargo_test.ps1
Using cargo
$ cargo +$(<RUST_NIGHTLY_VERSION) test
These binary overrides can be great for rapid iteration, but will hide errors like 4832 and 4834. To do a more authentic test:
- Build a new habitat package
- Upload to acceptance (or prod, but only in the unstable channel)
- Install new package and do normal testing
Changes to the exporters can be tested once the exporter package has been built locally. For example, to test changes to the Cloud Foundry exporter (core/hab-pkg-cfize
), first enter the studio and build a new package:
➤ hab studio enter
…
[1][default:/src:0]# build components/pkg-cfize
…
hab-pkg-cfize: Installed Path: /hab/pkgs/jbauman/hab-pkg-cfize/0.56.0-dev/20180410205025
Now your modifications are installed locally, under <HAB_ORIGIN>/hab-pkg-cfize
. You can run your new exporter with
[6][default:/src:0]# hab pkg exec $HAB_ORIGIN/hab-pkg-cfize hab-pkg-cfize --help
hab-pkg-export-cf 0.56.0-dev/20180410205025
…
Note that the version is updated, confirming you're running the new code. The old version is still accessible by running
[10][default:/src:1]# hab pkg export cf --help
hab-pkg-export-cf 0.55.0/20180321215151
…
See also Exporting Docker Images for Cloud Foundry with Habitat.
In order to test the Cloud Foundry exporter, you'll need a Cloud Foundry environment which supports Docker images. If you have one already, you can skip to step 4.
Note: Make sure to install PCF Dev on your base operating system. It uses VirtualBox as its virtualizer so you can't install that within a VM. Fortunately, PCF Dev is has support for Linux, Windows and Mac OS.
- Install the
cf
CLI. - Install VirtualBox 5.0+.
- Install PCF Dev. Note: running
cf dev start
for the first time requires a very large download (>7 GB as of this writing). - Build a
.hart
file and create a CF mapping.toml
. See Stark and Wayne's Running Habitat Apps on Cloud Foundry and Exporting to Cloud Foundry documentation for more details. - Run the exporter to add the
cf-
tagged image to Docker. docker push …
the image to Docker Hub.cf push …
the image from Docker Hub to add to the CF instance. Note: with PCF Dev this command must be run from the source directory of the project to correctly identify the necessary buildpack.
This one is a bit special. Technically hab-studio.sh is a shell script file and not binary. This also means that there is no need to build anything; set HAB_STUDIO_BINARY
to the path to a version of hab-studio.sh
within a habitat
checkout and it will be used. This override will also affect which versions of the files in studio/libexec are used. So if you want to test out changes to hab-studio-profile.sh or hab-studio-type-default.sh, make those changes in a checkout of the habitat
repo located at /path/to/habitat/repo
and set HAB_STUDIO_BINARY
to /path/to/habitat/repo/components/studio/bin/hab-studio.sh
. For example:
$ env HAB_STUDIO_BINARY=/path/to/habitat/repo/components/studio/bin/hab-studio.sh hab studio enter
Once inside the studio, the prompt will indicate that the override is in effect:
[1][HAB_STUDIO_BINARY][default:/src:0]#
These install instructions assume you want to develop, build, and run the various Habitat software components in a Linux environment. The Habitat core team suggests that you use our consistent development environment that we call the "devshell" as the easiest way to get started.
- Install Docker for Mac
- Checkout the source by running
git clone git@github.com:habitat-sh/habitat.git; cd habitat
- Run
make
to compile all Rust software components (this will take a while) - (Optional) Run
make test
if you want to run the tests. This will take a while.
Everything should come up green. Congratulations - you have a working Habitat development environment.
You can enter a devshell by running make shell
. This will drop you in a
Docker container at a Bash shell. The source code is mounted in under /src
,
meaning you can use common Rust workflows such as cd components/sup; cargo build
.
Note: The Makefile targets are documented. Run make help
to show the
output (this target requires perl
).
Optional: This project compiles and runs inside Docker containers so while
installing the Rust language isn't strictly necessary, you might want a local
copy of Rust on your workstation (some editors' language support require an
installed version). To install stable
Rust, run: curl -sSf https://sh.rustup.rs | sh
. Additionally, the project maintainers use the
nightly version of rustfmt
for code formatting. If you are submitting changes, please ensure that your
work has been run through a nightly version of rustfmt. The recommended way to
install it (assuming you have Rust installed as above), is to run
rustup component add --toolchain nightly rustfmt
and adding $HOME/.cargo/bin
to your PATH
. Actually running it would look something like cargo +nightly fmt --all
from the root of the project. The actual toolchain name can vary, depending on which
version of nightly rust you have installed.
Note2: While this Docker container will work well for getting started with Habitat development, you may want to consider using a VM as you start compiling Habitat components. To do this, create a VM with your preferred flavor of Linux and follow the appropriate instructions for that flavor below.
These instructions assume you want to develop, build, and run the various
Habitat software components in a macOS environment. Components other than
the hab
CLI itself aren't all supported on Mac, but it is sometimes useful
to run parts of the Habitat ecosystem without virtualization. We recommend
using another environment.
First clone the codebase and enter the directory:
git clone https://github.com/habitat-sh/habitat.git
cd habitat
Then, run the system preparation scripts and try to compile the project:
cp components/hab/install.sh /tmp/
sh support/mac/install_dev_0_mac_latest.sh
sh support/mac/install_dev_9_mac.sh
. ~/.profile
export PKG_CONFIG_PATH="/usr/local/opt/openssl/lib/pkgconfig:/usr/local/lib/pkgconfig:$PKG_CONFIG_PATH"
export IN_DOCKER=false
make
For the builds to find the openssl libraries, you will need to
set the PKG_CONFIG_PATH
environment variable as above before running cargo build
, cargo test
, etc. Additionally to use the Makefile on Mac and not have
the tasks execute in the Docker-based devshell (see above), you will need to
set IN_DOCKER=false
in your environment. If you use an environment switcher
such as direnv, you can set up the following in the root
of the git repository:
echo 'export PKG_CONFIG_PATH="/usr/local/opt/openssl/lib/pkgconfig:/usr/local/lib/pkgconfig:$PKG_CONFIG_PATH"' > .envrc
direnv allow
This can be used to build and install on older versions of Ubuntu where libsodium and czmq aren't available.
First clone the codebase and enter the directory:
git clone https://github.com/habitat-sh/habitat.git
cd habitat
Then, run the system preparation scripts and try to compile the project:
sh support/linux/install_dev_0_ubuntu_14.04.sh
sh support/linux/install_dev_9_linux.sh
. ~/.profile
make
These docs were tested with a Docker image, created as follows:
docker run --rm -it ubuntu:trusty bash
apt-get update && apt-get install -y sudo git-core
useradd -m -s /bin/bash -G sudo jdoe
echo jdoe:1234 | chpasswd
sudo su - jdoe
First clone the codebase and enter the directory:
git clone https://github.com/habitat-sh/habitat.git
cd habitat
Then, run the system preparation scripts and try to compile the project:
sh support/linux/install_dev_0_centos_7.sh
sh support/linux/install_dev_9_linux.sh
. ~/.profile
make
If you have issues with libsodium at runtime, ensure that you've set
LD_LIBRARY_PATH
and PKG_CONFIG_PATH
:
export LD_LIBRARY_PATH=/usr/local/lib
export PKG_CONFIG_PATH=/usr/local/lib/pkgconfig
These docs were tested with a Docker image, created as follows:
docker run --rm -it centos:7 bash
yum install -y sudo git
useradd -m -s /bin/bash -G wheel jdoe
echo jdoe:1234 | chpasswd
sudo su - jdoe
First clone the codebase and enter the directory:
git clone https://github.com/habitat-sh/habitat.git
cd habitat
Then, run the system preparation scripts and try to compile the project:
sh support/linux/install_dev_0_arch.sh
sh support/linux/install_dev_9_linux.sh
. ~/.profile
make
These docs were tested with a Docker image, created as follows:
docker run --rm -it greyltc/archlinux bash
pacman -Syy --noconfirm
pacman -S --noconfirm sudo git
echo "%wheel ALL=(ALL) ALL" > /etc/sudoers.d/01_wheel
useradd -m -s /bin/bash -G wheel jdoe
echo jdoe:1234 | chpasswd
sudo su - jdoe
While the available docker images can provide a convenient contributor onboarding experience, they may not prove ideal for extensive habitat development. Building habitat components is a disk intensive operation. Mounted file systems across network boundaries, even when confined to a single local host, can yield build times dramatically slower than building against a locally attached file system. Build times will be best on either a bare metal linux environment or a dedicated linux vm. Also note that when using vagrant to provision a vm for habitat development, it is advised that you do not put your habitat repo on a synced folder or at least do not use that folder as your build target.
The following directions clone the habitat repository into a local directory to realize faster build times; however, while faster, this approach carries the risk of losing your work if your VM should crash or get into an unrecoverable state. If developing in a VM, intermittently pushing a WIP to your Habitat fork can protect your progress.
In the root of the project is a Vagrantfile that provisions an Ubuntu environment for Habitat development:
vagrant up --provider virtualbox # See the Vagrantfile for additional providers and boxes
Feel free to use this file as a jumping-off point for customizing your own Habitat development environment.
vagrant ssh
sudo su -
git clone https://github.com/habitat-sh/habitat.git
cd habitat/components/builder-api
cargo test
These instructions are based on Windows 10 1607 (Anniversary update) or newer. Most of it will probably work downlevel.
All commands are in PowerShell unless otherwise stated. It is assumed that you
have git
installed and configured. Posh-Git is a handy PowerShell module for
making git
better in your PowerShell console (install-module posh-git
).
# Clone the Habitat source
git clone https://github.com/habitat-sh/habitat.git
cd habitat
./build.ps1 components/hab -configure