Skip to content
View othyn's full-sized avatar
🏳️‍🌈
Be kind, and have an adventure.
🏳️‍🌈
Be kind, and have an adventure.

Organizations

@KammaData

Block or report othyn

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
othyn/README.md

Ahoy-hoy! 👋

You'll find all sorts in here - to be slightly more coherent than /dev/urandom, that's the goal. I would say private or proprietary repo's are where the treasures are at, but depends on what you class as 'treasure', yarr... yarrr.

Be kind, and have an adventure. ♥️

Pinned Loading

  1. macos-auto-clicker macos-auto-clicker Public

    A simple auto clicker for macOS Big Sur, Monterey, Ventura, Sonoma and Sequoia.

    Swift 259 30

  2. github-folder-icon-macOS github-folder-icon-macOS Public

    GitHub folder icon for use in macOS. Made so that it fits nicely alongside all the others inside your home directory. Mac OS X 10.5 (Leopard) through to macOS 15 (Sequoia).

    Makefile 61

  3. go-calendar go-calendar Public

    A community driven auto-updating Pokémon GO events calendar that you can subscribe to in any calendar app on your phone/PC. Powered by Leek Duck.

    PHP 36 12

  4. Setting up a local only LLM (Qwen/Ll... Setting up a local only LLM (Qwen/Llama3/etc.) on macOS with Ollama, Continue and VSCode
    1
    # Setting up a local only LLM (Qwen/Llama3/etc.) on macOS with Ollama, Continue and VSCode
    2
    
                  
    3
    As with a lot of organisations, the idea of using LLM's is a reasonably frightning concept, as people freely hand over internal IP and sensitive comms to remote entities that are heavily data bound by nature. I know it was on our minds when deciding on LLM's and their role within the team and wider company. 6 months ago, I set out to explore what offerings were like in the self-hosted and/or OSS space, and if anything could be achieved locally. After using this setup since then, and after getting a lot of questions on it, I thought I might share some of the things I've come across and getting it all setup.
    4
    
                  
    5
    Que in [Ollama](https://ollama.com/) and [Continue](https://marketplace.visualstudio.com/items?itemName=Continue.continue). Ollama is an easy way to locally download, manage and run models. Its very familiar to Docker in its usuage, and can probably be most conceptually aligned with it in how it operates, think images = llm's. Continue is the other side of that, being the bridge/interconnect to allow what's in VSCode to talk with Ollama in a way that makes sense.
  5. Fix horrendously bad macOS (12.3.1 t... Fix horrendously bad macOS (12.3.1 tested) SMB (Samba) performance on Unraid
    1
    # Intro
    2
    
                  
    3
    Out of the box, my SMB performance on macOS 12.3.1 would top out at around 20MB/s in short ~5 second bursts, which was absolutely horrendous, slow to navigate in Finder and slugish to interact with.
    4
    
                  
    5
    Since making these changes, I now get sustained ~80-100MB/s+ and instant Finder navigation which is superb and how things should be out-of-the-box (OOTB)!
  6. How to disable default menu bar item... How to disable default menu bar items in Swift / SwiftUI for macOS
    1
    //
    2
    //  App.swift
    3
    //
    4
    //  Created by Ben Tindall on 30/03/2022.
    5
    //