Falcon 180B is a state of the-art AI model known for its superior text generation capabilities. Hosting it in a Docker container on an AWS EC2 instance ensures a reliable and scalable deployment.
Connect to Your EC2 Server
Before setting up Falcon 180B, establish a connection to your Amazon EC2 instance:
ssh -i "your-key.pem" ubuntu@your-ec2-ip
Once inside, update the system to ensure you have the latest packages:
sudo apt update && sudo apt upgrade -y
If Docker isn’t installed, set it up with:
sudo apt install docker.io -y
sudo systemctl start docker
sudo systemctl enable docker
Launch the Ollama Container
Deploy the Ollama runtime within a container to handle Falcon 180B:
docker run -d --name ollama -p 11434:11434 -v
ollama:/root/.ollama ollama/ollama
Install the Falcon 180B Model
Now, access the container and download Falcon 180B:
docker exec -it ollama /bin/bash
ollama pull falcon:180b
This process fetches all necessary dependencies.
Running Falcon 180B
Start the Model
Once the installation is complete, run the model:
ollama run falcon:180b
Try it out by entering a sample query:
>>> Summarize the latest advancements in AI.
For a more interactive experience, deploy a web UI:
docker run -d --name web-ui -p 3100:8080 -e
LLAMA_BASE_URL=http://<YOUR-EC2-IP>:11434 -v
open-webui:/app/backend/data --restart always
ghcr.io/open-webui/open-webui:main
Now, access http://<YOUR-EC2-IP>:3100 in your browser to interact visually.
Running Falcon 180B on AWS EC2 using Docker and Ollama provides a powerful, scalable and efficient AI deployment setup. Whether accessed via command line or web UI, this method ensures seamless operation with minimal setup efforts.
Ready to transform your business with our technology solutions? Contact Us today to Leverage Our AI/ML Expertise.