• 1 Post
  • 41 Comments
Joined 7 months ago
cake
Cake day: December 14th, 2023

help-circle











  • I just discovered how easy ollama and open webui are to set up so I’ve been using llama3 locally too, it was like 20 lines in docker compose, and although I’ve been using gpt3.5 on and off for a long time I’m much more comfortable using models run locally so I’ve been playing with it a lot more. It’s also cool being able to easily switch models at any point during a conversation. I have like 15 models downloaded, mostly 7b and a few 13b models and they all run fast enough on CPU and generate slightly slower than reading speed and only take ~15-30 seconds to start spitting out a response.

    Next I want to set up a vscode plugin so I can use my own locally run codegen models from within vscode.



  • Did you make sure to stop network manager too? I think disabling it tells it not to start it automatically but I think if it was already running it may have stayed up and maybe it brought the interface back up.

    That’s my only guess, if ip link shows it as down still then idk. NetworkManager also has its own Mac spoofing thing so you might have better success editing the properties of the network connection in NetworkManager and putting a new Mac in the cloned Mac address field. I’ve only used macchanger with netctl.