Download the worker script using a curl command or download the files through github.
Ensure you have the following installed and running on your system:
- Ollama: Download Ollama
Download this repository, extract the folder, and drag and drop infera-windows.exe to your desktop.
Double click on the infera-windows.exe icon on your desktop and the Infera node will boot up.
Please ensure that Ollama is running or infera-windows.exe will not boot up.
Download the shell script and place into your root directory and then open your terminal and copy and paste the following commands into your command line and press enter to install the Infera Node.
Note: The current Linux build is only for Linux + Intel chips, AMD chips will be added shortly
chmod +x ./infera-apple-m.sh
./infera-apple-m.sh
echo "alias init-infera='~/infera'" >> ~/.zshrc && source ~/.zshrc
chmod +x ./infera-linux-intel.sh
./infera-linux-intel.sh
To start the node, enter the following to your terminal and the node will automatically start
init-infera
To update Infera on your computer, enter the following command into your Terminal to remove and replace the depricated version of Infera.
rm -rf ~/infera
chmod +x ./infera-apple-m.sh
./infera-apple-m.sh
rm -rf ~/infera
chmod +x ./infera-linux-intel.sh
./infera-linux-intel.sh
To delete infera from your computer, enter the following command into your Terminal.
rm -rf ~/infera
While the node is running, you can install llama3:latest with CLI using the following terminal
command in a second terminal window.
curl -X POST http://localhost:11025/install_new_model \
-H "Content-Type: application/json" \
-d '{"install_model_name": "llama3:latest"}'
The node should now have llama3:latest installed and your node is now fully operational on the network.
llama3:latest can also be replace with the following models
gemma:latest
gemma2:latest
dolphin-mistral:latest
mistral:latest
llama3:latest
llama3.1:latest
llama2-uncensored:latest
Users can also user their browser and go to localhost:11025/docs to view the node API and see node stats and information.
Models can also be managed under the node API without having to use a terminal curl request.
Infera offers a Google Extension for users who wish to use a visual interface for accessing their hosted node.