• 0 Posts
  • 93 Comments
Joined 9 months ago
cake
Cake day: May 22nd, 2025

help-circle
  • It’s very content specific, what are you processing with the API?

    One of my little side projects right now is translating Russian fiction, specifically a genre over there called ‘boyar-anime’ which is essentially fantasy set in imperial russia. I do most my heavy translation using Anthropic Haiku which is very cheap and unlike the higher end models it tends to dumb down some of the more complex parts of Imperial Russian aristocracy so it’s more in line with similar fiction over here. When I take the source book, I chunk it down into small segments that I translate individually so I don’t get context bleed, then I mechanically process to find anything that didn’t translate very well. I combine roughly 40 of these weirdly translated segments into a jsonl file and submit the file through the API. OpenAI Batch API can accept up to 900k tokens, but you’ll wait close to 11 hours for something that large. 40 segments is around 30k tokens and that usually processes in a few mins to an hour depending.

    The jsonl file is essentially made up of smaller json blocks

    {
      "custom_id": "SEGMENT-NUM",
      "method": "POST",
      "url": "/v1/responses",
      "body": {
        "model": "gpt-5.3",
        "input": [
          {
            "role": "system",
            "content": [
              {
                "type": "input_text",
                "text": "You are a meticulous English language proofreader."
              }
            ]
          },
          {
            "role": "user",
            "content": [
              {
                "type": "input_text",
                "text": "PROMPT - SUBMITTED SEGMENT"
              }
            ]
          }
        ],
        "max_output_tokens": 8192
      }
    }
    

    I then setup polling to check back with the API every few mins, when the submitted queries are completed, I send more automatically until everything has been processed.



  • I don’t know what thinking profile your friend was using but asking ChatGPT that with the mixed tasks profile showed an almost immediate result with absolutely no thinking required.

    LLM’s are a tool, like with any tool there is a learning curve, and in my opinion the majority of “AI” users are unable to use the tool properly, and then get mad at the tool. Or like you, want to disparage the use of an LLM so they bait the LLM with tasks that it knows will fail or hallucinate on. To me that’s like blaming the table saw because it cut off your finger. Do the majority of people need a paid account? No.

    Are there people working in the Tech sector who use an LLM everyday, who have corporate accounts and paid accounts at home for their own projects: absolutely. I know a large number of them, most are Lemmy users as well. But because there is so much negativity from the open source crowd, all these engineers are afraid to discuss all the ways it makes our lives easier. So we get a disproportionate amount of negativity. I’m getting to a point where the amount of AI shit posting on here is like the amount of vegan shit posting on Reddit. And just as stupid.


  • Opus is heavily throttled outside enterprise tiers. I was regularly blowing through weekly usage limits by Tuesday using Opus. 5.3 on the higher thinking profiles match or exceed Opus capabilities, and I have yet to hit a single limitation.

    If I need to process via API I will run tests against Anthropic Haiku or Sonnet before trying Gpt5-mini, If I need to use 5.3, and what I’m doing isn’t time critical I’ll use batch processing. Smaller token batches complete very quickly, often in under 2 hours. And at a 50% discount, provides serious cost savings.


  • This being Lemmy and AI shit posting a hobby of everyone on here. I’ve had excellent results with AI. I have weird complicated health issues and in my search for ways not to die early from these issues AI is a helpful tool.

    Should you trust AI? of course not but having used Gemini, then Claude and now ChatGPT I think how you interact with the AI makes the difference. I know what my issues are, and when I’ve found a study that supports an idea I want to discuss with my doctor I will usually first discuss it with AI. The Canadian healthcare landscape is such that my doctor is limited to a 15min appt, part of a very large hospital associated practice with a large patient load. He uses AI to summarize our conversation, and to look up things I bring up in the appointment. I use AI to preplan my appointment, help me bring supporting documentation or bullet points my doctor can then use to diagnose.

    AI is not a doctor, but it helps both me and my doctor in this situation we find ourselves in. If I didn’t have access to my doctor, and had to deal with the American healthcare system I could see myself turning to AI for more than support. AI has never steered me wrong, both Gemini and Claude have heavy guardrails in place to make it clear that AI is not a doctor, and AI should not be a trusted source for medical advice. I’m not sure about ChatGPT as I generally ask that any guardrails be suppressed before discussing medical topics. When I began using ChatGPT I clearly outlined my health issues and so far it remembers that context, and I haven’t received hallucinated diagnoses. YMMV.







  • Well I’ve done most of those except for using an immutable distro, and quite honestly have zero interest in running one.

    Just this past year I’ve setup a media server using a bunch of servers I nabbed from work, did unraid because I’ve been toying with the unraid idea for years. Learned and fell in love with Docker. I avoided it like the plague for the better part of a decade. Now I run damn near everything in docker containers. Started using home assistant quite recently actually, my wife got annoyed with the heavily sanitized results she was getting with Gemini for Google Home. So I setup M5Stack ATOM Lite ESP32’s around the house acting as microphones that I then use with home assistant to ferry the query to the Anthropic API and the result to the nearest Google Home device. It’s a work in progress, but happy with the results


  • Bought a deep fryer a couple years back, once you’re done with it, and let it cool down. You can turn a knob and it will filter itself and empty into a storage container. Once the oil has gone bad, I have a Home Depot bucket with a lid that I dump it into, once that is full I take it to the dump. Once the filtered bits are dumped in the compost, everything else goes in the dishwasher.

    I fry in peanut oil, I have 2 air fryers but some things need to be deep fried.



  • Honestly, I didn’t know what to do either. I’m a big iron sysadmin but I’m old now, so I’m mostly relegated to management. We have young guys who do all the cloud and virtualization crap and they make me look like the dinosaur that I am.

    But like many companies we had AI forced on us, so when I didn’t know what filesystem to use on my array I asked Claude. Claude knows everything about unRaid. Every roadblock I hit was answered so thoroughly I hit my usage limit and had to make a note on my profile that Claude was not to make documentation unless I damn well asked for it.

    I know AI is not a popular topic but it’s honestly made my life better. I’ve been running Linux since forever but still when an update breaks Arch, and reverting to a snapshot doesn’t fix it. I usually just reinstall. Now I don’t have to feel stupid asking younger guys how to fix an issue. I just type the issue into Claude, drop into TTY and I have all the info needed to fix it within minutes.

    Just yesterday one of my coworkers calls me up and says Santa was good to him and gives me a Radeon that was leaps ahead of what I was running so I checked the CachyOS wiki, thought I was prepared to switch from Nvidia to Radeon and the second the switch was made everything went to shit. Claude to the rescue with all the commands to purge my system of nividia garbage and reinstall the Radeon versions. He also ran me through adding my old nividia card to unraid so I can use it for my vm’s.